m-polignano-uniba
commited on
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,103 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- gsarti/clean_mc4_it
|
4 |
+
- Chat-Error/wizard_alpaca_dolly_orca
|
5 |
+
- jondurbin/truthy-dpo-v0.1
|
6 |
+
- mlabonne/orpo-dpo-mix-40k
|
7 |
+
base_model: meta-llama/Meta-Llama-3-8B-Instruct
|
8 |
+
model_creator: Marco Polignano - SWAP Research Group
|
9 |
+
language:
|
10 |
+
- en
|
11 |
+
- it
|
12 |
+
metrics:
|
13 |
+
- accuracy
|
14 |
+
pipeline_tag: text-generation
|
15 |
+
tags:
|
16 |
+
- facebook
|
17 |
+
- meta
|
18 |
+
- pythorch
|
19 |
+
- llama
|
20 |
+
- llama-3
|
21 |
+
- llamantino
|
22 |
+
library_name: transformers
|
23 |
+
license: llama3
|
24 |
+
---
|
25 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/5df8bb21da6d0311fd3d540f/cZoZdwQOPdQsnQmDXHcSn.png" alt="llamantino3_anita" border="0" width="800px">
|
26 |
+
<hr>
|
27 |
+
<!--<img src="https://i.ibb.co/6mHSRm3/llamantino53.jpg" width="200"/>-->
|
28 |
+
|
29 |
+
<p style="text-align:justify;"><b>LLaMAntino-3-ANITA-8B-Inst-DPO-ITA</b> is a model of the <a href="https://huggingface.co/swap-uniba"><b>LLaMAntino</b></a> - <i>Large Language Models family</i>.
|
30 |
+
The model is an instruction-tuned version of <a href="https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct"><b>Meta-Llama-3-8b-instruct</b></a> (a fine-tuned <b>LLaMA 3 model</b>).
|
31 |
+
This model version aims to be the a <b>Multilingual Model</b> ๐ (EN ๐บ๐ธ + ITA๐ฎ๐น) to further fine-tuning on Specific Tasks in Italian.</p>
|
32 |
+
|
33 |
+
|
34 |
+
The ๐**ANITA project**๐ *(**A**dvanced **N**atural-based interaction for the **ITA**lian language)*
|
35 |
+
wants to provide Italian NLP researchers with an improved model the for Italian Language ๐ฎ๐น use cases.
|
36 |
+
|
37 |
+
<hr>
|
38 |
+
|
39 |
+
## Model Details
|
40 |
+
|
41 |
+
<img src="https://static.vecteezy.com/system/resources/previews/016/833/880/large_2x/github-logo-git-hub-icon-with-text-on-white-background-free-vector.jpg" width="200"> [https://github.com/marcopoli/LLaMAntino-3-ANITA](https://github.com/marcopoli/LLaMAntino-3-ANITA)<br>
|
42 |
+
|
43 |
+
<br>
|
44 |
+
- [**Full Model: LaMAntino-3-ANITA-8B-Inst-DPO-ITA**](https://huggingface.co/m-polignano-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA)
|
45 |
+
- LLaMA.cpp - **F16 model**
|
46 |
+
- LLaMA.cpp - **Q8_0 model**
|
47 |
+
- LLaMA.cpp - **Q4_K_M model**
|
48 |
+
- LLaMA.cpp - **Q2_K model**
|
49 |
+
|
50 |
+
<hr>
|
51 |
+
|
52 |
+
## Specifications
|
53 |
+
|
54 |
+
- **Model developers**: <br><a href="https://marcopoli.github.io/">Ph.D. Marco Polignano</a> - University of Bari Aldo Moro, Italy <br> <a href="https://huggingface.co/swap-uniba">SWAP Research Group</a> <br>
|
55 |
+
- **Variations**: The model release has been **supervised fine-tuning (SFT)** using **QLoRA** 4bit, on two instruction-based datasets. **DPO** approach over the *jondurbin/truthy-dpo-v0.1* dataset is used to align with human preferences for helpfulness and safety.
|
56 |
+
- **Input**: Models input text only.
|
57 |
+
- **Language**: Multilingual ๐ + Italian ๐ฎ๐น
|
58 |
+
- **Output**: Models generate text and code only.
|
59 |
+
- **Model Architecture**: *Llama 3 architecture*.
|
60 |
+
- **Context length**: 8K, 8192.
|
61 |
+
- **Library Used**: [LLaMA.cpp](https://github.com/ggerganov/llama.cpp)
|
62 |
+
|
63 |
+
<hr>
|
64 |
+
|
65 |
+
### Prompt Template
|
66 |
+
```
|
67 |
+
<|start_header_id|>system<|end_header_id|>
|
68 |
+
|
69 |
+
{ SYS Prompt }<|eot_id|><|start_header_id|>user<|end_header_id|>
|
70 |
+
|
71 |
+
{ USER Prompt }<|eot_id|><|start_header_id|>assistant<|end_header_id|>
|
72 |
+
|
73 |
+
{ ASSIST Prompt }<|eot_id|>
|
74 |
+
````
|
75 |
+
|
76 |
+
<hr>
|
77 |
+
|
78 |
+
## LLaMA.cpp
|
79 |
+
|
80 |
+
<img src="https://user-images.githubusercontent.com/1991296/230134379-7181e485-c521-4d23-a0d6-f7b3b61ba524.png" width="200px" align="center" />
|
81 |
+
|
82 |
+
[LLaMA.cpp](https://github.com/ggerganov/llama.cpp), a great tool that helps us easily Quantize your model in **GGUF format**.
|
83 |
+
|
84 |
+
## Citation instructions
|
85 |
+
```bibtex
|
86 |
+
@misc{basile2023llamantino,
|
87 |
+
title={LLaMAntino: LLaMA 2 Models for Effective Text Generation in Italian Language},
|
88 |
+
author={Pierpaolo Basile and Elio Musacchio and Marco Polignano and Lucia Siciliani and Giuseppe Fiameni and Giovanni Semeraro},
|
89 |
+
year={2023},
|
90 |
+
eprint={2312.09993},
|
91 |
+
archivePrefix={arXiv},
|
92 |
+
primaryClass={cs.CL}
|
93 |
+
}
|
94 |
+
```
|
95 |
+
|
96 |
+
```bibtex
|
97 |
+
@article{llama3modelcard,
|
98 |
+
title={Llama 3 Model Card},
|
99 |
+
author={AI@Meta},
|
100 |
+
year={2024},
|
101 |
+
url = {https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md}
|
102 |
+
}
|
103 |
+
```
|