Update README.md
Browse files
README.md
CHANGED
@@ -11,14 +11,14 @@ license: llama3.3
|
|
11 |
---
|
12 |
# about
|
13 |
|
14 |
-
The Teaz series is my third attempt at making merges, after the Kostume and Kermes series.
|
15 |
|
16 |
This time, the goal was to make a smart model with a low perplexity, in accordance to the principles of the Kermes series, but with a merge of 3 merged models like on the kostume series.
|
17 |
|
18 |
Huihui's abliterated models were used:
|
19 |
- Llama 3.3 70b as the pivot of the first/main model.
|
20 |
- Nemotron 3.1 70b and Deepseek R1 Distill 70b as the pillars.
|
21 |
-
- and Tulu 3 70b as the
|
22 |
|
23 |
Bingo again. I hit 3.45 ppl512 wikieng, 62+ or ARC-C, and 82+ on ARC-E. Absolute top of the class for L3.x 70b, like Kermes is for L3 3.2 3b.
|
24 |
|
|
|
11 |
---
|
12 |
# about
|
13 |
|
14 |
+
The Teaz series is my third attempt at making merges, this time on L3.x 70b, after the L3.2 3b Kostume and Kermes series.
|
15 |
|
16 |
This time, the goal was to make a smart model with a low perplexity, in accordance to the principles of the Kermes series, but with a merge of 3 merged models like on the kostume series.
|
17 |
|
18 |
Huihui's abliterated models were used:
|
19 |
- Llama 3.3 70b as the pivot of the first/main model.
|
20 |
- Nemotron 3.1 70b and Deepseek R1 Distill 70b as the pillars.
|
21 |
+
- and Tulu 3 70b as the backer of the 2nd and 3rd models.
|
22 |
|
23 |
Bingo again. I hit 3.45 ppl512 wikieng, 62+ or ARC-C, and 82+ on ARC-E. Absolute top of the class for L3.x 70b, like Kermes is for L3 3.2 3b.
|
24 |
|