Konan-LLM
Collection
1 item
β’
Updated
Konan-LLM-OND, a large language model from Konan Technology Inc., is based on Qwen3-4B-Base. It has been specifically optimized for the Korean language through vocabulary expansion, continual pre-training, and instruction tuning to enhance performance and efficiency.
Model | Model size | Korean | English | ||||
---|---|---|---|---|---|---|---|
KMMLU | HRM8K | Ko-IFEval | MMLU | GSM8K | IFEval | ||
Konan-LLM-OND | 4.0B | 50.6 | 46.4 | 68.4 | 68.8 | 86.8 | 73.3 |
EXAONE-3.5-2.4B-Instruct | 2.4B | 44.2 | 31.8 | 60.5 | 59.1 | 81.5 | 77.7 |
kanana-1.5-2.1b-instruct-2505 | 2.1B | 32.7 | 27.2 | 56.0 | 52.9 | 68.8 | 64.6 |
Midm-2.0-Mini-Instruct | 2.3B | 42.4 | 36.2 | 66.8 | 57.4 | 74.8 | 68.3 |
Qwen3-4B(w/o reasoning) | 4.0B | 0.0(*) | 37.5 | 68.4 | 29.4(*) | 83.9 | 80.0 |
gemma-3-4b-it | 4.3B | 38.7 | 32.7 | 69.2 | 59.1 | 82.2 | 78.3 |
Model | Model size | Korean | English | ||||
---|---|---|---|---|---|---|---|
KMMLU | HRM8K | Ko-IFEval | MMLU | GSM8K | IFEval | ||
Konan-LLM-OND | 4.0B | 50.6 | 46.4 | 68.4 | 68.8 | 86.8 | 73.3 |
A.X-4.0-Light | 7.2B | 55.3 | 44.6 | 71.5 | 70.6 | 87.3 | 81.3 |
EXAONE-3.5-7.8B-Instruct | 7.8B | 48.0 | 39.3 | 66.8 | 66.8 | 91.4 | 79.9 |
kanana-1.5-8b-instruct-2505 | 8.0B | 40.4 | 35.5 | 71.1 | 63.1 | 79.3 | 76.8 |
Midm-2.0-Base-Instruct | 11.5B | 54.2 | 46.0 | 75.0 | 70.2 | 88.9 | 79.7 |
Qwen3-8B(w/o reasoning) | 8.1B | 0.0(*) | 40.0 | 70.9 | 7.4(*) | 84.0 | 82.8 |
Note:
All benchmarks were executed using the following standardized environment.
lm-evaluation-harness v0.4.9
vLLM v0.9.1
on a single NVIDIA GPU.chat_completions
API, and scores were computed solely from the generated responses.Benchmark | Scoring Method | Few-shot |
---|---|---|
KMMLU | exact_match |
5-shot |
HRM8K | mean of hrm8k_gsm8k , hrm8k_ksm , hrm8k_math , hrm8k_mmmlu , hrm8k_omni_math |
5-shot |
Ko-IFEval | mean of prompt_level_strict_acc , inst_level_strict_acc , prompt_level_loose_acc , inst_level_loose_acc |
0-shot |
MMLU | exact_match |
5-shot |
GSM8K | exact_match & flexible-extract |
5-shot |
IFEval | mean of prompt_level_strict_acc , inst_level_strict_acc , prompt_level_loose_acc , inst_level_loose_acc |
0-shot |
Konan-LLM-OND is supported in transformers v4.52.0
and later.
pip install transformers>=4.52.0
The code example below shows you how to get the model to generate content based on given inputs.
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_name = "konantech/Konan-LLM-OND"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.bfloat16,
device_map="auto",
)
model.eval()
tokenizer = AutoTokenizer.from_pretrained(model_name)
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "λνλ―Όκ΅ μλλ?"}
]
input_ids = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device)
with torch.no_grad():
output = model.generate(
input_ids,
max_new_tokens=64,
do_sample=False,
)
len_input_prompt = len(input_ids[0])
response = tokenizer.decode(output[0][len_input_prompt:], skip_special_tokens=True)
print(response)
# λνλ―Όκ΅ μλλ μμΈμ
λλ€.
@misc{Konan-LLM-OND-2025,
author = {Konan Technology Inc.},
title = {Konan-LLM-OND},
year = {2025},
url = {https://huggingface.co/konantech/Konan-LLM-OND}
}
Base model
Qwen/Qwen3-4B-Base