siwon-mini-instruct-0626

This model is a fine-tuned version of microsoft/Phi-4-mini-instruct, adapted for Korean instruction-based tasks. The tuning was focused on enhancing Korean performance through supervised fine-tuning with Korean instruction datasets.


πŸ”§ Token Adjustments

The original model used the same token ID (199999) for multiple special tokens such as BOS, EOS, PAD, and UNK. This caused confusion in instruction-following tasks. We fixed this by remapping the token IDs as follows:

Token Type Original ID Fixed ID
BOS 199999 199999
EOS 199999 200020
PAD 199999 200029
UNK 199999 200030

These changes ensure proper differentiation and functioning of special tokens during generation and training.


πŸ—¨οΈ Chat Template

The chat template was updated accordingly to support multi-turn conversation formatting in the Korean context:

{% for message in messages %}
{% if message['role'] == 'system' and 'tools' in message and message['tools'] is not none %}
{{ '<|' + message['role'] + '|>' + message['content'] + '<|tool|>' + message['tools'] + '<|/tool|>' + '<|end|>' }}
{% else %}
{{ '<|' + message['role'] + '|>' + message['content'] + '<|end|>' }}
{% endif %}
{% endfor %}
{% if add_generation_prompt %}{{ '<|assistant|>' }}{% endif %}

πŸ§ͺ Inference with Transformers

Below is an example of how to load and use the model with the adjusted tokenizer, token IDs, and custom prompt template.

Note: This model uses a custom chat_template and updated special token IDs:

  • <|end|> β†’ 200020 (EOS)
  • <|dummy_85|> β†’ 200029 (PAD)
  • Γ―ΒΏΒ½ β†’ 200030 (UNK)
from transformers import AutoTokenizer, AutoModelForCausalLM

model_path = "madcows/siwon-mini-instruct-0626"

model = AutoModelForCausalLM.from_pretrained(
    model_path,
    device_map="auto",
    torch_dtype=torch.bfloat16,
    trust_remote_code=True
)

tokenizer = AutoTokenizer.from_pretrained(
    model_path,
    trust_remote_code=True,
)

messages = [
        {"role": "system", "content": "You are a helpful assistant."},
        {"role": "user", "content": "μ•ˆλ…•ν•˜μ„Έμš”."},
    ]

inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)

output = model.generate(
    **inputs,
    max_new_tokens=2048,
    # do_sample=True, # Optional
    # top_p=0.95, # Optional
    # temperature=0.6, # Optional
    # repetition_penalty=1.1, # Optional
)

response = tokenizer.decode(output[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True)
print(response)

πŸ“Š Model Performance Comparison

Performance scores across three Korean language benchmarks (KMMLU, ko_best, pawsx_ko).

Model KMMLU (0-shot) ko_best (5-shot) pawsx_ko
Phi-4-mini-instruct 0.3161 0.6341 0.5300
kanana-1.5-2.1b-instruct-2505 0.1577 0.7165 0.5070
EXAONE-3.5-2.4B-Instruct 0.3071 0.6496 0.5655
siwon-mini-instruct-0626 0.3387 0.5576 0.5485

πŸ“Œ Caution

  • Commercial use is strictly prohibited.

  • This model is intended for research and educational use only.

  • Redistribution or use in commercial products or services is not allowed.

✍️ Acknowledgments

  • Base model: microsoft/Phi-4-mini-instruct
  • Special thanks to the open-source community for instruction-tuning resources and Korean language corpora.

πŸ™ Feedback & Contributions

We welcome any feedback to improve the model’s performance, usability, and alignment with Korean instruction tasks. If you encounter any issues or have suggestions, please feel free to open an issue on the Hugging Face model page.

Your input is greatly appreciated and will help us enhance the model further.

Downloads last month
40
Safetensors
Model size
3.84B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for madcows/siwon-mini-instruct-0626

Finetuned
(34)
this model