A.X 3.1 Light
A.X 3.1 Light is a lightweight large language model (LLM) optimized for Korean understanding and enterprise deployment, fully independently developed by SKT.
Downloads 411
Release Time : 7/10/2025
Model Overview
A.X 3.1 Light is an efficient multilingual large language model that particularly focuses on Korean capabilities and supports long context processing and multilingual tasks.
Model Features
Pure Korean sovereign AI
Fully independently developed by SKT, trained using its own GPU infrastructure and high-quality datasets.
Excellent Korean capabilities
Performs excellently in the Korean language evaluation benchmarks KMMLU and KoBALT - 700, surpassing other Korean-specific models.
Efficient token usage
When processing Korean input, it uses 33% fewer tokens than GPT - 4o, making it more cost - effective and computationally efficient.
Long context processing capabilities
Supports long context processing of up to 32,768 tokens.
Model Capabilities
Korean text generation
Multilingual translation
Instruction following
Mathematical calculation
Code generation
Use Cases
Translation
Korean - English translation
Translate Korean text into English or other languages.
High - quality translation results, suitable for enterprise documents and multilingual communication.
Enterprise deployment
Customer support
Used in automated customer support systems to handle Korean queries.
Efficiently respond to customer needs and reduce labor costs.
🚀 A.X 3.1 Light
A.X 3.1 Light is a lightweight large language model (LLM) optimized for Korean - language understanding and enterprise deployment. It offers high - efficiency and cost - effective compute usage, with strong performance in multiple benchmarks.
✨ Features
- Authentic Korean Sovereign AI: Trained on a high - quality multilingual dataset curated in - house using SKT’s proprietary GPU infrastructure.
- Highly Efficient Multilingual LLM: Demonstrates superior performance among open - source Korean LLMs, despite a relatively compact training size of 1.65 trillion tokens.
- Superior Korean Proficiency: Achieved a score of 61.7 on the [KMMLU](https://huggingface.co/datasets/HAERAE - HUB/KMMLU), outperforming other Korean - specified models.
- Deep Korean Understanding: Obtained 27.43 on the [KoBALT - 700](https://huggingface.co/datasets/snunlp/KoBALT - 700), outperforming other Korean - specialized models.
- Efficient Token Usage: Requires approximately 33% fewer tokens than GPT - 4o to process equivalent Korean inputs.
- Long - Context Handling: Supports up to 32,768 tokens.
🔧 Technical Details
Model Architecture Specs
Property | Details |
---|---|
Model | A.X 3.1 Light |
# Params | 7B |
# Layers | 32 |
# KV - Heads | 32 |
Hidden Dim | 4096 |
FFN Dim | 10880 |
High - Quality Data Pipeline & Strategic Mixture
- A training dataset of 20 trillion tokens from diverse domains was collected and curated.
- The entire dataset was processed through SKT’s proprietary data pipeline, including synthetic data generation and comprehensive quality filtering.
- A total of 1.65 trillion tokens from a Korean - focused multilingual corpus were used for training A.X 3.1 Light.
Pareto - Optimal Compute Efficiency
A.X 3.1 Light has 5 to 6 times lower computational cost compared to models with similar performance levels. Rigorous data curation and two - stage training with STEM - focused data enabled competitive performance at reduced FLOPs.
📚 Documentation
Benchmark Results
Benchmarks | Sub - Benchmark | A.X 3.1 Light | Kanana - 1.5 - 8B | EXAONE - 3.5 - 7.8B | Qwen2.5 - 7B | Qwen3 - 8B (w/o reasoning) |
---|---|---|---|---|---|---|
Knowledge | KMMLU | 61.70 | 48.28 | 53.76 | 49.56 | 63.53 |
CLIcK | 71.22 | 61.30 | 64.11 | 58.30 | 63.31 | |
KoBALT | 27.43 | 23.14 | 21.71 | 21.57 | 26.57 | |
MMLU | 66.95 | 68.82 | 72.20 | 75.40 | 82.89 | |
General | Ko - MT - Bench | 78.56 | 76.30 | 81.06 | 61.31 | 64.06 |
MT - Bench | 74.38 | 77.60 | 83.50 | 79.37 | 65.69 | |
Instruction Following |
Ko - IFEval | 70.04 | 69.96 | 65.01 | 60.73 | 73.39 |
IFEval | 79.86 | 80.11 | 82.61 | 76.73 | 85.38 | |
Math | HRM8K | 41.70 | 30.87 | 31.88 | 35.13 | 52.50 |
MATH | 70.14 | 59.28 | 63.20 | 65.58 | 71.48 | |
Code | HumanEval+ | 73.78 | 76.83 | 76.83 | 74.39 | 77.44 |
MBPP+ | 61.64 | 67.99 | 64.29 | 68.50 | 62.17 |
📦 Installation
with HuggingFace Transformers
transformers>=4.46.0
or the latest version is required to useskt/A.X - 3.1 - Light
pip install transformers>=4.46.0
with vLLM
vllm>=v0.6.4.post1
or the latest version is required to use tool - use feature
pip install vllm>=v0.6.4.post1
# if you don't want to activate tool - use feature, just commenting out below vLLM option
VLLM_OPTION="--enable-auto-tool-choice --tool-call-parser hermes"
vllm serve skt/A.X-3.1-Light $VLLM_OPTION
💻 Usage Examples
Basic Usage
with HuggingFace Transformers
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "skt/A.X-3.1-Light"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16,
device_map="auto",
)
model.eval()
tokenizer = AutoTokenizer.from_pretrained(model_name)
messages = [
{"role": "system", "content": "당신은 사용자가 제공하는 영어 문장들을 한국어로 번역하는 AI 전문가입니다."},
{"role": "user", "content": "The first human went into space and orbited the Earth on April 12, 1961."},
]
input_ids = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
with torch.no_grad():
output = model.generate(
input_ids,
max_new_tokens=128,
do_sample=False,
)
len_input_prompt = len(input_ids[0])
response = tokenizer.decode(output[0][len_input_prompt:], skip_special_tokens=True)
print(response)
# Output:
# 1961년 4월 12일, 최초의 인간이 우주에 나가 지구를 궤도를 돌았습니다.
with vLLM
from openai import OpenAI
def call(messages, model):
completion = client.chat.completions.create(
model=model,
messages=messages,
)
print(completion.choices[0].message)
client = OpenAI(
base_url="http://localhost:8000/v1",
api_key="api_key"
)
model = "skt/A.X-3.1-Light"
messages = [{"role": "user", "content": "에어컨 여름철 적정 온도는? 한줄로 답변해줘"}]
call(messages, model)
# Output:
# 에어컨 여름철 적정 온도는 24~26도입니다.
messages = [{"role": "user", "content": "What is the appropriate temperature for air conditioning in summer? Respond in a single sentence."}]
call(messages, model)
# Output:
# The appropriate temperature for air conditioning in summer is generally set between 24 to 26°C for optimal comfort and energy efficiency.
Advanced Usage (Examples for tool - use)
from openai import OpenAI
def call(messages, model):
completion = client.chat.completions.create(
model=model,
messages=messages,
tools=tools
)
print(completion.choices[0].message)
client = OpenAI(
base_url="http://localhost:8000/v1",
api_key="api_key"
)
model = "skt/A.X-3.1-Light"
calculate_discount = {
"type": "function",
"function": {
"name": "calculate_discount",
"description": "원가격과 할인율(퍼센트 단위)을 입력받아 할인된 가격을계산한다.",
"parameters": {
"type": "object",
"properties": {
"original_price": {
"type": "number",
"description": "상품의 원래 가격"
},
"discount_percentage": {
"type": "number",
"description": "적용할 할인율"
}
},
"required": ["original_price", "discount_percentage"]
}
}
}
get_exchange_rate = {
"type": "function",
"function": {
"name": "get_exchange_rate",
"description": "두 통화 간의 환율을 가져온다.",
"parameters": {
"type": "object",
"properties": {
"base_currency": {
"type": "string",
"description": "The currency to convert from."
},
"target_currency": {
"type": "string",
"description": "The currency to convert to."
}
},
"required": ["base_currency", "target_currency"]
}
}
}
tools = [calculate_discount, get_exchange_rate]
### Slot filling ###
messages = [{"role": "user", "content": "우리가 뭘 사야되는데 원가가 57600원인데 직원할인 받으면 얼마야?"}]
call(messages, model)
# Output:
# ChatCompletionMessage(content='직원 할인을 적용하기 위해서는 할인율을 알 수 있어야 합니다. 할인율을 알려주실 수 있나요?', role='assistant', function_call=None, tool_calls=[], reasoning_content=None)
### Function calling ###
messages = [
{"role": "user", "content": "우리가 뭘 사야되는데 원가가 57600원인데 직원할인 받으면 얼마야?"},
{"role": "assistant", "content": "직원 할인을 적용하기 위해서는 할인율을 알 수 있어야 합니다. 할인율을 알려주실 수 있나요?"},
{"role": "user", "content": "15% 할인 받을 수 있어."},
]
call(messages, model)
# Output:
# ChatCompletionMessage(content=None, role='assistant', function_call=None, tool_calls=[ChatCompletionMessageToolCall(id='chatcmpl-tool-3ebf11847364450daf363039db80cc50', function=Function(arguments='{"original_price": 57600, "discount_percentage": 15}', name='calculate_discount'), type='function')], reasoning_content=None)
### Completion ###
messages = [
{"role": "user", "content": "우리가 뭘 사야되는데 원가가 57600원인데 직원할인 받으면 얼마야?"},
{"role": "assistant", "content": ""},
{"role": "user", "content": "15% 할인 받을 수 있어."},
{"role": "tool", "tool_call_id": "random_id", "name": "calculate_discount", "content": "{\"original_price\": 57600, \"discount_percentage\": 15, \"discounted_price\": 48960.0}"}
]
call(messages, model)
# Output:
# ChatCompletionMessage(content='57,600원의 상품에 15% 할인을 적용하면, 할인된 가격은 48,960원입니다.', role='assistant', function_call=None, tool_calls=[], reasoning_content=None)
📄 License
The A.X 3.1 Light
model is licensed under Apache License 2.0
.
Citation
@article{SKTAdotX3.1Light,
title={A.X 3.1 Light},
author={SKT AI Model Lab},
year={2025},
url={https://huggingface.co/skt/A.X-3.1-Light}
}
Contact
- Business & Partnership Contact: a.x@sk.com
Phi 2 GGUF
Other
Phi-2 is a small yet powerful language model developed by Microsoft, featuring 2.7 billion parameters, focusing on efficient inference and high-quality text generation.
Large Language Model Supports Multiple Languages
P
TheBloke
41.5M
205
Roberta Large
MIT
A large English language model pre-trained with masked language modeling objectives, using improved BERT training methods
Large Language Model English
R
FacebookAI
19.4M
212
Distilbert Base Uncased
Apache-2.0
DistilBERT is a distilled version of the BERT base model, maintaining similar performance while being more lightweight and efficient, suitable for natural language processing tasks such as sequence classification and token classification.
Large Language Model English
D
distilbert
11.1M
669
Llama 3.1 8B Instruct GGUF
Meta Llama 3.1 8B Instruct is a multilingual large language model optimized for multilingual dialogue use cases, excelling in common industry benchmarks.
Large Language Model English
L
modularai
9.7M
4
Xlm Roberta Base
MIT
XLM-RoBERTa is a multilingual model pretrained on 2.5TB of filtered CommonCrawl data across 100 languages, using masked language modeling as the training objective.
Large Language Model Supports Multiple Languages
X
FacebookAI
9.6M
664
Roberta Base
MIT
An English pre-trained model based on Transformer architecture, trained on massive text through masked language modeling objectives, supporting text feature extraction and downstream task fine-tuning
Large Language Model English
R
FacebookAI
9.3M
488
Opt 125m
Other
OPT is an open pre-trained Transformer language model suite released by Meta AI, with parameter sizes ranging from 125 million to 175 billion, designed to match the performance of the GPT-3 series while promoting open research in large-scale language models.
Large Language Model English
O
facebook
6.3M
198
1
A pretrained model based on the transformers library, suitable for various NLP tasks
Large Language Model
Transformers

1
unslothai
6.2M
1
Llama 3.1 8B Instruct
Llama 3.1 is Meta's multilingual large language model series, featuring 8B, 70B, and 405B parameter scales, supporting 8 languages and code generation, with optimized multilingual dialogue scenarios.
Large Language Model
Transformers Supports Multiple Languages

L
meta-llama
5.7M
3,898
T5 Base
Apache-2.0
The T5 Base Version is a text-to-text Transformer model developed by Google with 220 million parameters, supporting multilingual NLP tasks.
Large Language Model Supports Multiple Languages
T
google-t5
5.4M
702
Featured Recommended AI Models