Text Generation
Transformers
Safetensors
lola_v1
custom_code
neo-nlp-dev commited on
Commit
1c9069d
1 Parent(s): a970019

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -3
README.md CHANGED
@@ -5,11 +5,14 @@ datasets:
5
  - uonlp/CulturaX
6
  ---
7
 
8
- # Model Card for Model ID
9
-
10
  <!-- Provide a quick summary of what the model is/does. -->
11
  # LOLA &mdash; An Open-Source Massively Multilingual Large Language Model
12
 
 
 
 
 
 
13
 
14
  ## Model Description
15
 
@@ -21,7 +24,7 @@ datasets:
21
  - **License:** CC BY 4.0 (https://creativecommons.org/licenses/by/4.0/)
22
  - **Repository:** https://github.com/dice-group/LOLA
23
 
24
- <sub>* The number of parameters a model utilizes per token (ref: [Du et al, 2022](https://arxiv.org/abs/2112.06905)). This distinction is crucial for understanding the efficiency and performance of MoE models.</sub>
25
 
26
  ## How to Get Started with the Model
27
 
 
5
  - uonlp/CulturaX
6
  ---
7
 
 
 
8
  <!-- Provide a quick summary of what the model is/does. -->
9
  # LOLA &mdash; An Open-Source Massively Multilingual Large Language Model
10
 
11
+ ## Abstract
12
+
13
+ This paper presents LOLA, a massively multilingual large language model trained on more than 160 languages using a sparse Mixture-of-Experts Transformer architecture. Our architectural and implementation choices address the challenge of harnessing linguistic diversity while maintaining efficiency and avoiding the common pitfalls of multilinguality. Our analysis of the evaluation results shows competitive performance in natural language generation and understanding tasks. Additionally, we demonstrate how the learned expert-routing mechanism exploits implicit phylogenetic linguistic patterns to potentially alleviate the curse of multilinguality. We provide an in-depth look at the training process, an analysis of the datasets, and a balanced exploration of the model's strengths and limitations. As an open-source model, LOLA promotes reproducibility and serves as a robust foundation for future research. Our findings enable the development of compute-efficient multilingual models with strong, scalable performance across languages.
14
+
15
+ Paper: https://arxiv.org/abs/2409.11272
16
 
17
  ## Model Description
18
 
 
24
  - **License:** CC BY 4.0 (https://creativecommons.org/licenses/by/4.0/)
25
  - **Repository:** https://github.com/dice-group/LOLA
26
 
27
+ <sub>* The number of parameters a model utilizes per token (ref: [Fedus et al, 2022](https://arxiv.org/abs/2101.03961); [Du et al, 2022](https://arxiv.org/abs/2112.06905)). This distinction is crucial for understanding the efficiency and performance of MoE models.</sub>
28
 
29
  ## How to Get Started with the Model
30