haoranxu commited on
Commit
8e68401
·
verified ·
1 Parent(s): 13e39b0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -1
README.md CHANGED
@@ -3,7 +3,16 @@ license: mit
3
  ---
4
  **[ALMA-R](https://arxiv.org/abs/2401.08417)** builds upon [ALMA models](https://arxiv.org/abs/2309.11674), with further LoRA fine-tuning with our proposed **Contrastive Preference Optimization (CPO)** as opposed to the Supervised Fine-tuning used in ALMA. CPO fine-tuning requires our [triplet preference data](https://huggingface.co/datasets/haoranxu/ALMA-R-Preference) for preference learning. ALMA-R now can matches or even exceeds GPT-4 or WMT winners!
5
 
6
-
 
 
 
 
 
 
 
 
 
7
  # Download ALMA(-R) Models and Dataset 🚀
8
 
9
  We release six translation models presented in the paper:
 
3
  ---
4
  **[ALMA-R](https://arxiv.org/abs/2401.08417)** builds upon [ALMA models](https://arxiv.org/abs/2309.11674), with further LoRA fine-tuning with our proposed **Contrastive Preference Optimization (CPO)** as opposed to the Supervised Fine-tuning used in ALMA. CPO fine-tuning requires our [triplet preference data](https://huggingface.co/datasets/haoranxu/ALMA-R-Preference) for preference learning. ALMA-R now can matches or even exceeds GPT-4 or WMT winners!
5
 
6
+ ```
7
+ @misc{xu2024contrastive,
8
+ title={Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation},
9
+ author={Haoran Xu and Amr Sharaf and Yunmo Chen and Weiting Tan and Lingfeng Shen and Benjamin Van Durme and Kenton Murray and Young Jin Kim},
10
+ year={2024},
11
+ eprint={2401.08417},
12
+ archivePrefix={arXiv},
13
+ primaryClass={cs.CL}
14
+ }
15
+ ```
16
  # Download ALMA(-R) Models and Dataset 🚀
17
 
18
  We release six translation models presented in the paper: