Update README.md
Browse files
README.md
CHANGED
@@ -60,8 +60,18 @@ base_model:
|
|
60 |
---
|
61 |
|
62 |
|
63 |
-
X-ALMA builds upon [ALMA-R](https://arxiv.org/pdf/2401.08417) by expanding support from 6 to 50 languages. It utilizes a plug-and-play architecture with language-specific modules, complemented by a carefully designed training recipe. This release includes the **X-ALMA pre-trained base model**.
|
64 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
65 |
X-ALMA-13B-Pretrain is pre-trained on 50 languages: en,da,nl,de,is,no,sv,af,ca,ro,gl,it,pt,es,bg,mk,sr,uk,ru,id,ms,th,vi,mg,fr,hu,el,cs,pl,lt,lv,ka,zh,ja,ko,fi,et,gu,hi,mr,ne,ur,az,kk,ky,tr,uz,ar,he,fa.
|
66 |
|
67 |
All X-ALMA checkpoints are released at huggingface:
|
|
|
60 |
---
|
61 |
|
62 |
|
63 |
+
[X-ALMA](https://arxiv.org/pdf/2410.03115) builds upon [ALMA-R](https://arxiv.org/pdf/2401.08417) by expanding support from 6 to 50 languages. It utilizes a plug-and-play architecture with language-specific modules, complemented by a carefully designed training recipe. This release includes the **X-ALMA pre-trained base model**.
|
64 |
+
```
|
65 |
+
@misc{xu2024xalmaplugplay,
|
66 |
+
title={X-ALMA: Plug & Play Modules and Adaptive Rejection for Quality Translation at Scale},
|
67 |
+
author={Haoran Xu and Kenton Murray and Philipp Koehn and Hieu Hoang and Akiko Eriguchi and Huda Khayrallah},
|
68 |
+
year={2024},
|
69 |
+
eprint={2410.03115},
|
70 |
+
archivePrefix={arXiv},
|
71 |
+
primaryClass={cs.CL},
|
72 |
+
url={https://arxiv.org/abs/2410.03115},
|
73 |
+
}
|
74 |
+
```
|
75 |
X-ALMA-13B-Pretrain is pre-trained on 50 languages: en,da,nl,de,is,no,sv,af,ca,ro,gl,it,pt,es,bg,mk,sr,uk,ru,id,ms,th,vi,mg,fr,hu,el,cs,pl,lt,lv,ka,zh,ja,ko,fi,et,gu,hi,mr,ne,ur,az,kk,ky,tr,uz,ar,he,fa.
|
76 |
|
77 |
All X-ALMA checkpoints are released at huggingface:
|