Model Information

preview

Kumru-2B is the lightweight, open-source version of Kumru LLM, developed for Turkish from scratch by VNGRS.

  • It is pre-trained on a cleaned, deduplicated corpora of 500 GB for 300B tokens, and supervised fine-tuned on 1M examples.
  • It comes with a modern tokenizer developed for Turkish, supporting code, math and chat template.
  • Kumru has a native context length of 8,192 tokens by default.
  • This is the Base version, which is only pre-trained, not instruct fine-tuned.
  • Instruct fine-tuned version is here

Try to demo of 7B version here.

Use

from transformers import AutoModelForCausalLM, AutoTokenizer

model_name = "vngrs-ai/Kumru-2B-Base"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", device_map="auto")

def complete_text(starting_text):
    tokenized_text = tokenizer.encode_plus(starting_text, return_tensors = 'pt', return_token_type_ids=False,
                                           max_length = model.config.max_position_embeddings, truncation = True).to(model.device)
    generated_tokens = model.generate(**tokenized_text, 
                                      max_length = 512,
                                      repetition_penalty = 1.15,
                                      no_repeat_ngram_size = 5
                                     )
    generated_text = tokenizer.decode(generated_tokens.cpu().detach().numpy().reshape(-1).tolist(), skip_special_tokens=True)

    return generated_text

starting_text = "Efes antik kenti"
generated_text = complete_text(starting_text)
print(generated_text)

Evaluation Results

Both Kumru-7B and Kumru-2B are evaluated on Cetvel benchmark.

preview

We observe that Kumru overall surpasses significantly larger models such as LLaMA-3.3–70B, Gemma-3–27B, Qwen-2–72B and Aya-32B. It excels at tasks related to the nuances of the Turkish language, such as grammatical error correction and text summarization.

Tokenizer Efficiency

Kumru tokenizer is a modern BPE tokenizer with a vocabulary size of 50,176, pre-tokenization regex and a chat template.

preview

Other open-source models spend between 38% to 98% more tokens than Kumru while still having larger vocabulary sizes. This means Kumru can represent more texts in its context length and process faster and cheaper. Although the native context length of Kumru is 8,192, its effective context length can be considered between 1128 and 1618, compared to other multilingual models out there. This shows the efficiency of having a native Turkish tokenizer in terms of representation power, speed and cost.

Citation

@misc{turker2025kumru,
  title={Kumru},
  author={Turker, Meliksah and Ari, Erdi and Han, Aydin},
  year={2025},
  url={https://huggingface.co/vngrs-ai/Kumru-2B-Base}
}
Downloads last month
931
Safetensors
Model size
2B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for vngrs-ai/Kumru-2B-Base

Adapters
2 models
Finetunes
1 model

Dataset used to train vngrs-ai/Kumru-2B-Base

Space using vngrs-ai/Kumru-2B-Base 1

Collection including vngrs-ai/Kumru-2B-Base