Tri-1.8B Translation

We release Tri-1.8B Translation, a lightweight multilingual translation model from Trillion Labs.

Tri-1.8B Translate is trained through pretraining and supervised fine-tuning (SFT), and was distilled from our larger Tri-21B model to preserve strong translation quality in a much smaller, deployment-friendly 1.8B parameter model. It supports all translation directions among English, Korean, Japanese, and Chinese.


✨ Highlights

  • Compact & efficient: ~1.8B parameters, easy to deploy.
  • Multilingual: Fully bidirectional translation across EN ↔ KO ↔ JA ↔ ZH.
  • Research-ready: Ideal for experimentation and domain fine-tuning.

πŸ”§ Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "trillionlabs/Tri-1.8B-Translation"

tok = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto")

prompt = "Translate English to Korean: 'We look forward to working with you again.' <ko>"

inputs = tok(prompt, return_tensors="pt").to(model.device)
out = model.generate(**inputs, max_new_tokens=128)
print(tok.decode(out[0], skip_special_tokens=True))
Downloads last month
181
Safetensors
Model size
1.81B params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support