This model is an Instruction-Tuned version of Llama 3.2 400M Amharic.

How to use

Chat Format

Given the nature of the training data, the phi-2 instruct model is best suited for prompts using the chat format as follows. You can provide the prompt as a question with a generic template as follows:

<|im_start|>user
αŒ₯ያቄ?<|im_end|>
<|im_start|>assistant

For example:

<|im_start|>user
αˆΆαˆ΅α‰΅ α‹¨αŠ ααˆͺካ αˆ€αŒˆαˆ«α‰΅ αŒ₯α‰€αˆ΅αˆαŠ<|im_end|>
<|im_start|>assistant

where the model generates the text after <|im_start|>assistant .

Sample inference code

First, you need to install the latest version of transformers

pip install -Uq transformers

You can use this model directly with a pipeline for text generation:

from transformers import pipeline

llama3_am = pipeline(
    "text-generation",
    model="rasyosef/Llama-3.2-400M-Amharic-Instruct",
    device_map="auto"
  )

messages = [{"role": "user", "content": "αˆ°αˆ‹αˆ"}]
llama3_am(messages, max_new_tokens=128, repetition_penalty=1.1, return_full_text=False)

Output:

[{'generated_text': 'αˆ°αˆ‹αˆ! α‹›αˆ¬ αŠ₯αŠ•α‹΄α‰΅ αˆαˆ¨α‹³α‹Žα‰΅ αŠ₯α‰½αˆ‹αˆˆαˆ? 😊'}]

Generate a Story

messages = [{"role": "user", "content": "ሡለ αŒ…α‰₯αŠ“ αŠ αŠ•α‰ αˆ³ α‰°αˆ¨α‰΅ αŠ•αŒˆαˆ¨αŠ"}]
llama3_am(messages, max_new_tokens=128, repetition_penalty=1.1, return_full_text=False)

Output:

[{'generated_text': 'α‰ αŠ₯α‹«αˆ± ኦαˆͺጎ α‹¨α‰°α‰°αˆ¨αŠ¨ α‹˜αŒ αŠ αŒ…α‰¦α‰½αŠ“ αŠ αŠ•α‹΅ αŠ αŠ•α‰ αˆ³ αŠ α‹°αŠ• α‹ˆαŒ₯ተው አሡር αˆ‹αˆžα‰½αŠ“ α‰ αˆ¬α‹Žα‰½αŠ• αŠ α‹΅αŠα‹ ያዙፑፑ αŠ αŠ•α‰ αˆ³α‹αˆ β€œαŠ₯αŠ” α‹˜αŒ αŠ™αŠ• α‹ˆαˆ΅αŒ„ αŠ₯αŠ“αŠ•α‰° αŠ αŠ•α‹±αŠ• α‰΅α‹ˆαˆ΅α‹³αˆ‹α‰½αˆα‘α‘β€ αˆ²αˆ‹α‰Έα‹ αŒ…α‰¦α‰Ήαˆ β€œα‹­αˆ… αŠ₯αŠ•α‹΄α‰΅ α‹­αˆ†αŠ“αˆ? αŠ₯αŠ› α‹˜αŒ αŠ αˆ†αŠαŠ• ሳለ αŠ αŠ•α‰° α‹˜αŒ αŠ™αŠ• α‹ˆαˆ΅α‹°αˆ… ለαŠ₯αŠ› αŠ αŠ•α‹΅ αˆ˜αˆ΅αŒ α‰΅αˆ… αˆšα‹›αŠ“α‹Š αŠα‹αŠ•?” αŠ αˆ‰α‰΅α‘α‘ αŠ αŠ•α‰ αˆ³α‹αˆ β€œαŠ α‹Ž αŠ₯αŠ” αŠ αŠ•α‹΅ ነኝፑፑ α‹˜αŒ αŠ™αŠ• αˆ΅α‹ˆαˆ΅α‹΅ አሡር αŠ₯αŠ•αˆ†αŠ“αˆˆαŠ•α‘α‘ αŠ₯αŠ“αŠ•α‰°αˆ α‹˜αŒ αŠ αŠ“α‰½αˆα‘α‘ αŠ αŠ•α‹·αŠ• αˆ‹αˆ αˆ΅α‰΅αŒ¨αˆαˆ© አሡር α‰΅αˆ†αŠ“αˆ‹α‰½αˆα‘α‘β€ αŠ αˆ‹α‰Έα‹α‘α‘ αŒ…α‰¦α‰Ήαˆ α‹¨αŠ αŠ•α‰ αˆ³α‹αŠ• አሡፈαˆͺ αŠ α‹­αŠ–α‰½ α‰ α‰°αˆ˜αˆˆαŠ¨α‰± αŒŠα‹œ αˆ΅αˆˆαˆαˆ©α‰΅ α‹ˆα‹° αŠ α‰£α‰³α‰Έα‹ α‰ αˆ˜αˆ„α‹΅ αŠ αŠ•α‰ αˆ³α‹ ለαŠ₯ነርሱ αŠ αŠ•α‹΅ ሰαŒ₯ቷቸው ለራሱ α‹˜αŒ αŠ αŠ₯αŠ•α‹°α‹ˆαˆ°α‹°αŠ“ α‹­αŠΈα‹αˆ αŠ αŠ•α‰ αˆ³α‹ αŠ₯αŠ› α‹˜αŒ αŠ'}]

Generate a Poem

messages = [{"role": "user", "content": "ሡለ α‹­α‰…αˆ­α‰³ ግαŒ₯ም ጻፍልኝ"}]
llama3_am(messages, max_new_tokens=128, repetition_penalty=1.1, return_full_text=False)

Output:

[{'generated_text': 'α‹­α‰…αˆ­α‰³ α‹­α‰…αˆ­α‰³\nα‹­α‰…αˆ­α‰³ αŠ₯αŠ“α‰΄\nαŠ α‹°α‰£α‰£α‹­ αˆ˜αˆ€αˆ αˆ‹α‰†αˆαŠ©α‰΅ αŠ•ααŒˆα‰΄\nሳመሽ αˆŒα‰΅αˆ½ አልፎኝ αˆ΅α‹αˆ α‰€αŠ•αˆ½ አልፎኝ\nጠገα‰₯ኩ αŠ₯ልሻለሁ αˆ¨αˆ€α‰₯ሽ α‰°αˆ­αŽαŠ\nα‹­α‰…αˆ­α‰³ αŠ α‰£α‰΄\nαŠ α‹°α‰£α‰£α‹­ αˆ˜αˆ€αˆ αˆ‹α‰†αˆαŠ©α‰΅ αˆ΅αˆ΅α‰΄\nαŠ¨αˆαŒ…αŠα‰΅ α‹ˆαŠ” αˆˆαˆ‹α‰€α‹ αˆžαŒˆαˆ΅αˆ…\nαŠ¨αˆžα‰΄ αŒˆα‹˜αˆ αˆαŒ„ αŠα‰΅ αˆ›αŠαˆ΅αˆ…\nα‹­α‰…αˆ­α‰³ α‹ˆαŠ•α‹΅αˆœ\nαŠ α‹°α‰£α‰£α‹­ αˆ˜αˆ€αˆ αˆ‹α‰†αˆαŠ©α‰΅ αˆ…αˆ˜αˆœ\nαŠ¨αŠ α‰» αˆ˜αŠ•αŒˆα‹³α‰½αŠ• αˆˆαŠ“α‰…αŠ©α‰΅ αˆ›αŒ£α‰΅αˆ…\nሡጠኝ αˆ›αˆˆα‰΅ ከα‰₯α‹Άαˆ… ሡαŒ₯ል αˆˆαˆ›αŠ•αˆ³α‰΅αˆ…\nα‹­α‰…αˆ­α‰³ αŠ₯αˆ…α‰΄\nαŠ α‹°α‰£α‰£α‹­ αˆ˜αˆ€αˆ αˆ‹α‰†αˆαŠ©α‰΅ αˆ½αŠ•αˆα‰΄\nአαŒ₯αˆ¬αŠ• አαŒ₯αˆ¬αŠ• ሡል αˆˆαŒ£αˆαŠ©α‰΅ αŠ α‰ α‰£αˆ½\nα‰ αˆ³α‰„'}]
Downloads last month
15
Safetensors
Model size
413M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for rasyosef/Llama-3.2-400M-Amharic-Instruct-Poems-Stories-Wikipedia

Finetuned
(8)
this model

Space using rasyosef/Llama-3.2-400M-Amharic-Instruct-Poems-Stories-Wikipedia 1

Collection including rasyosef/Llama-3.2-400M-Amharic-Instruct-Poems-Stories-Wikipedia