Muennighoff commited on
Commit
0b4f30a
1 Parent(s): 5da805c
Files changed (1) hide show
  1. README.md +10 -3
README.md CHANGED
@@ -694,7 +694,7 @@ model-index:
694
  > We present BLOOMZ & mT0, a family of models capable of following human instructions in dozens of languages zero-shot. We finetune BLOOM & mT5 pretrained multilingual language models on our crosslingual task mixture (xP3) and find our resulting models capable of crosslingual generalization to unseen tasks & languages.
695
 
696
  - **Repository:** [bigscience-workshop/xmtf](https://github.com/bigscience-workshop/xmtf)
697
- - **Paper:** [TODO]
698
  - **Point of Contact:** [Niklas Muennighoff](mailto:[email protected])
699
  - **Languages:** Refer to [mc4](https://huggingface.co/datasets/mc4) for pretraining & [xP3](https://huggingface.co/bigscience/xP3) for finetuning language proportions. It understands both pretraining & finetuning languages.
700
  - **BLOOMZ & mT0 Model Family:**
@@ -887,9 +887,16 @@ print(tokenizer.decode(outputs[0]))
887
 
888
  # Evaluation
889
 
890
- We refer to Table 7 from our paper [TODO LINK] & [bigscience/evaluation-results](https://huggingface.co/datasets/bigscience/evaluation-results) for zero-shot results on unseen tasks. The sidebar reports zero-shot performance of the best prompt per dataset config.
891
 
892
  # Citation
893
  ```bibtex
894
- TODO
 
 
 
 
 
 
 
895
  ```
 
694
  > We present BLOOMZ & mT0, a family of models capable of following human instructions in dozens of languages zero-shot. We finetune BLOOM & mT5 pretrained multilingual language models on our crosslingual task mixture (xP3) and find our resulting models capable of crosslingual generalization to unseen tasks & languages.
695
 
696
  - **Repository:** [bigscience-workshop/xmtf](https://github.com/bigscience-workshop/xmtf)
697
+ - **Paper:** [Crosslingual Generalization through Multitask Finetuning](https://arxiv.org/abs/2211.01786)
698
  - **Point of Contact:** [Niklas Muennighoff](mailto:[email protected])
699
  - **Languages:** Refer to [mc4](https://huggingface.co/datasets/mc4) for pretraining & [xP3](https://huggingface.co/bigscience/xP3) for finetuning language proportions. It understands both pretraining & finetuning languages.
700
  - **BLOOMZ & mT0 Model Family:**
 
887
 
888
  # Evaluation
889
 
890
+ We refer to Table 7 from our [paper](https://arxiv.org/abs/2211.01786) & [bigscience/evaluation-results](https://huggingface.co/datasets/bigscience/evaluation-results) for zero-shot results on unseen tasks. The sidebar reports zero-shot performance of the best prompt per dataset config.
891
 
892
  # Citation
893
  ```bibtex
894
+ @misc{muennighoff2022crosslingual,
895
+ title={Crosslingual Generalization through Multitask Finetuning},
896
+ author={Niklas Muennighoff and Thomas Wang and Lintang Sutawika and Adam Roberts and Stella Biderman and Teven Le Scao and M Saiful Bari and Sheng Shen and Zheng-Xin Yong and Hailey Schoelkopf and Xiangru Tang and Dragomir Radev and Alham Fikri Aji and Khalid Almubarak and Samuel Albanie and Zaid Alyafeai and Albert Webson and Edward Raff and Colin Raffel},
897
+ year={2022},
898
+ eprint={2211.01786},
899
+ archivePrefix={arXiv},
900
+ primaryClass={cs.CL}
901
+ }
902
  ```