This model is an Instruction-Tuned version of Llama 3.2 400M Amharic.

How to use

Chat Format

Given the nature of the training data, the phi-2 instruct model is best suited for prompts using the chat format as follows. You can provide the prompt as a question with a generic template as follows:

<|im_start|>user
αŒ₯ያቄ?<|im_end|>
<|im_start|>assistant

For example:

<|im_start|>user
αˆΆαˆ΅α‰΅ α‹¨αŠ ααˆͺካ αˆ€αŒˆαˆ«α‰΅ αŒ₯α‰€αˆ΅αˆαŠ<|im_end|>
<|im_start|>assistant

where the model generates the text after <|im_start|>assistant .

Sample inference code

First, you need to install the latest version of transformers

pip install -Uq transformers

You can use this model directly with a pipeline for text generation:

from transformers import pipeline

llama3_am = pipeline(
    "text-generation",
    model="rasyosef/Llama-3.2-400M-Amharic-Instruct",
    device_map="auto"
  )

messages = [{"role": "user", "content": "αˆΆαˆ΅α‰΅ α‹¨αŠ ααˆͺካ αˆ€αŒˆαˆ«α‰΅ αŒ₯α‰€αˆ΅αˆαŠ"}]
llama3_am(messages, max_new_tokens=128, repetition_penalty=1.1, return_full_text=False)

Output:

[{'generated_text': '1. ግα‰₯ፅ 2. αŠ“α‹­αŒ„αˆͺα‹« 3. αŒ‹αŠ“'}]
Downloads last month
136
Safetensors
Model size
413M params
Tensor type
F32
Β·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for rasyosef/Llama-3.2-400M-Amharic-Instruct

Finetuned
(2)
this model

Collection including rasyosef/Llama-3.2-400M-Amharic-Instruct