File size: 1,132 Bytes
1a6436a
 
 
 
 
 
 
ad92d5a
1a6436a
ad92d5a
1a6436a
547cf26
 
 
 
 
 
b33bc10
547cf26
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
---
library_name: transformers
tags:
- mergekit
- merge

---
# arco ultra

arco ultra is an improvement over [WizardLM-Evol-V2-Unfiltered](https://huggingface.co./trollek/danube2-1.8b-WizardLM-Evol-V2-Unfiltered) which at the time of writting is a state-of-the-art 1.8 billion parameters mistral language model. The model was first merged with another sota model ([Synthia-v1.3](https://huggingface.co./trollek/danube2-1.8b-Synthia-v1.3)) in order to get a balanced, performant model.

Post-merging was used just like with palmer models (merge, finetune later).

#### benchmarks

zero-shot evaluations performed on current sota ~0.5b models.

| Parameters | Model                          | MMLU  | ARC | HellaSwag | PIQA   | Winogrande | Average |
| -----------|--------------------------------|-------|-------|-----------|--------|------------|---------|
| 0.5b       | arco                           |26.17|37.29|62.88|74.37|62.27|52.60|
| 1.8B       | wizard                         |40.79|40.87|71.85|**78.02**|64.33| 59.17|
| 1.8B       | niro                           |**41.75**|**40.96**|**72.07**|77.97|**65.51**|**59.65**|