metadata
library_name: transformers
tags:
- mergekit
- merge
arco ultra
arco ultra is an improvement over WizardLM-Evol-V2-Unfiltered which at the time of writting is a state-of-the-art 1.8 billion parameters mistral language model. The model was first merged with another sota model (Synthia-v1.3) in order to get a balanced, performant model.
Post-merging was used just like with palmer models (merge, finetune later).
benchmarks
zero-shot evaluations performed on current sota ~0.5b models.
Parameters | Model | MMLU | ARC | HellaSwag | PIQA | Winogrande | Average |
---|---|---|---|---|---|---|---|
0.5b | arco | 26.17 | 37.29 | 62.88 | 74.37 | 62.27 | 52.60 |
1.8B | wizard | 40.79 | 40.87 | 71.85 | 78.02 | 64.33 | 59.17 |
1.8B | niro | 41.75 | 40.96 | 72.07 | 77.97 | 65.51 | 59.65 |