Nexesenex commited on
Commit
3780957
·
verified ·
1 Parent(s): b2f9064

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -11,14 +11,14 @@ license: llama3.3
11
  ---
12
  # about
13
 
14
- The Teaz series is my third attempt at making merges, after the Kostume and Kermes series.
15
 
16
  This time, the goal was to make a smart model with a low perplexity, in accordance to the principles of the Kermes series, but with a merge of 3 merged models like on the kostume series.
17
 
18
  Huihui's abliterated models were used:
19
  - Llama 3.3 70b as the pivot of the first/main model.
20
  - Nemotron 3.1 70b and Deepseek R1 Distill 70b as the pillars.
21
- - and Tulu 3 70b as the backers of the 2nd and 3rd models.
22
 
23
  Bingo again. I hit 3.45 ppl512 wikieng, 62+ or ARC-C, and 82+ on ARC-E. Absolute top of the class for L3.x 70b, like Kermes is for L3 3.2 3b.
24
 
 
11
  ---
12
  # about
13
 
14
+ The Teaz series is my third attempt at making merges, this time on L3.x 70b, after the L3.2 3b Kostume and Kermes series.
15
 
16
  This time, the goal was to make a smart model with a low perplexity, in accordance to the principles of the Kermes series, but with a merge of 3 merged models like on the kostume series.
17
 
18
  Huihui's abliterated models were used:
19
  - Llama 3.3 70b as the pivot of the first/main model.
20
  - Nemotron 3.1 70b and Deepseek R1 Distill 70b as the pillars.
21
+ - and Tulu 3 70b as the backer of the 2nd and 3rd models.
22
 
23
  Bingo again. I hit 3.45 ppl512 wikieng, 62+ or ARC-C, and 82+ on ARC-E. Absolute top of the class for L3.x 70b, like Kermes is for L3 3.2 3b.
24