Text Generation
Transformers
Safetensors
llama
conversational
text-generation-inference
Inference Endpoints
haoranxu commited on
Commit
c57c245
1 Parent(s): 0c92966

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -2
README.md CHANGED
@@ -60,8 +60,18 @@ base_model:
60
  ---
61
 
62
 
63
- X-ALMA builds upon [ALMA-R](https://arxiv.org/pdf/2401.08417) by expanding support from 6 to 50 languages. It utilizes a plug-and-play architecture with language-specific modules, complemented by a carefully designed training recipe. This release includes the **X-ALMA pre-trained base model**.
64
-
 
 
 
 
 
 
 
 
 
 
65
  X-ALMA-13B-Pretrain is pre-trained on 50 languages: en,da,nl,de,is,no,sv,af,ca,ro,gl,it,pt,es,bg,mk,sr,uk,ru,id,ms,th,vi,mg,fr,hu,el,cs,pl,lt,lv,ka,zh,ja,ko,fi,et,gu,hi,mr,ne,ur,az,kk,ky,tr,uz,ar,he,fa.
66
 
67
  All X-ALMA checkpoints are released at huggingface:
 
60
  ---
61
 
62
 
63
+ [X-ALMA](https://arxiv.org/pdf/2410.03115) builds upon [ALMA-R](https://arxiv.org/pdf/2401.08417) by expanding support from 6 to 50 languages. It utilizes a plug-and-play architecture with language-specific modules, complemented by a carefully designed training recipe. This release includes the **X-ALMA pre-trained base model**.
64
+ ```
65
+ @misc{xu2024xalmaplugplay,
66
+ title={X-ALMA: Plug & Play Modules and Adaptive Rejection for Quality Translation at Scale},
67
+ author={Haoran Xu and Kenton Murray and Philipp Koehn and Hieu Hoang and Akiko Eriguchi and Huda Khayrallah},
68
+ year={2024},
69
+ eprint={2410.03115},
70
+ archivePrefix={arXiv},
71
+ primaryClass={cs.CL},
72
+ url={https://arxiv.org/abs/2410.03115},
73
+ }
74
+ ```
75
  X-ALMA-13B-Pretrain is pre-trained on 50 languages: en,da,nl,de,is,no,sv,af,ca,ro,gl,it,pt,es,bg,mk,sr,uk,ru,id,ms,th,vi,mg,fr,hu,el,cs,pl,lt,lv,ka,zh,ja,ko,fi,et,gu,hi,mr,ne,ur,az,kk,ky,tr,uz,ar,he,fa.
76
 
77
  All X-ALMA checkpoints are released at huggingface: