Llama-3.2-Amharic-1B

This model is a version of Meta's Llama-3.2-1B decoder transformer model that was continuously pretrained on an Amharic text corpus.

  • 16k new amharic tokens were added to the Llama 3.2 tokenizer and the embdedding layer of the model was resized accordingly.
  • The model was then trained on 300 million tokens of Amharic text.
  • This is a base model. The Amharic instruction following version is Llama-3.2-1B-Amharic-Instruct

How to use

First, you need to install the latest version of transformers

pip install -Uq transformers

You can use this model directly with a pipeline for text generation:

from transformers import pipeline

llama_am = pipeline(
    "text-generation",
    model="rasyosef/Llama-3.2-1B-Amharic",
    device_map="auto"
  )

prompt = "በኢንግሊዝ ፕሪምየር ሊግ"
llama_am(
    prompt,
    max_new_tokens=128,
    temperature=0.3,
    do_sample=True,
    top_k=8,
    top_p=0.8,
    repetition_penalty=1.05
  )

Output:

[{'generated_text': 'በኢንግሊዝ ፕሪምየር ሊግ የ2017/18 የውድድር ዘመን ላይ ተሳታፊ የሆነው ሊቨርፑል ትናንት ምሽት 3 :45 ላይ ከዌስትሀም ዩናይትድ ጋር ባደረገው ጨዋታ በ2 ለ 1 ውጤት ተሸንፏል ።'}]
Downloads last month
117
Safetensors
Model size
1.27B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for rasyosef/Llama-3.2-1B-Amharic

Finetuned
(273)
this model
Finetunes
1 model
Quantizations
1 model

Collection including rasyosef/Llama-3.2-1B-Amharic