eval_name
stringlengths 12
111
| Precision
stringclasses 3
values | Type
stringclasses 6
values | T
stringclasses 6
values | Weight type
stringclasses 2
values | Architecture
stringclasses 52
values | Model
stringlengths 355
689
| fullname
stringlengths 4
102
| Model sha
stringlengths 0
40
| Average ⬆️
float64 1.03
52
| Hub License
stringclasses 26
values | Hub ❤️
int64 0
5.9k
| #Params (B)
int64 -1
140
| Available on the hub
bool 2
classes | MoE
bool 2
classes | Flagged
bool 2
classes | Chat Template
bool 2
classes | CO₂ cost (kg)
float64 0.03
107
| IFEval Raw
float64 0
0.9
| IFEval
float64 0
90
| BBH Raw
float64 0.27
0.75
| BBH
float64 0.81
63.5
| MATH Lvl 5 Raw
float64 0
0.51
| MATH Lvl 5
float64 0
50.7
| GPQA Raw
float64 0.22
0.44
| GPQA
float64 0
24.9
| MUSR Raw
float64 0.29
0.6
| MUSR
float64 0
38.5
| MMLU-PRO Raw
float64 0.1
0.73
| MMLU-PRO
float64 0
70
| Merged
bool 2
classes | Official Providers
bool 2
classes | Upload To Hub Date
stringclasses 424
values | Submission Date
stringclasses 169
values | Generation
int64 0
10
| Base Model
stringlengths 4
102
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
DreadPoor_BaeZel-8B-LINEAR_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/BaeZel-8B-LINEAR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/BaeZel-8B-LINEAR</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__BaeZel-8B-LINEAR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/BaeZel-8B-LINEAR | 1deac3287de191794c50543d69d523f43654a803 | 30.296459 | apache-2.0 | 1 | 8 | true | false | false | true | 0.665069 | 0.737792 | 73.779239 | 0.54638 | 35.535376 | 0.178248 | 17.824773 | 0.321309 | 9.50783 | 0.422708 | 13.338542 | 0.386137 | 31.792996 | true | false | 2024-11-08 | 2024-11-08 | 1 | DreadPoor/BaeZel-8B-LINEAR (Merge) |
DreadPoor_Condensed_Milk-8B-Model_Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Condensed_Milk-8B-Model_Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Condensed_Milk-8B-Model_Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Condensed_Milk-8B-Model_Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Condensed_Milk-8B-Model_Stock | 6e5b73099b9d5a794c9c744c4c5c158b1feb8916 | 30.070583 | apache-2.0 | 1 | 8 | true | false | false | true | 0.654664 | 0.753629 | 75.362926 | 0.543486 | 35.12062 | 0.173716 | 17.371601 | 0.321309 | 9.50783 | 0.41601 | 11.101302 | 0.387633 | 31.95922 | true | false | 2024-11-27 | 2024-11-27 | 1 | DreadPoor/Condensed_Milk-8B-Model_Stock (Merge) |
DreadPoor_CoolerCoder-8B-LINEAR_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/CoolerCoder-8B-LINEAR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/CoolerCoder-8B-LINEAR</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__CoolerCoder-8B-LINEAR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/CoolerCoder-8B-LINEAR | db14b0fa821b0b6b07802111fd19ba722344a32b | 19.148011 | 0 | 8 | false | false | false | true | 1.444021 | 0.451929 | 45.192866 | 0.47615 | 26.365383 | 0.061934 | 6.193353 | 0.290268 | 5.369128 | 0.396354 | 7.777604 | 0.315908 | 23.989731 | false | false | 2024-11-20 | 0 | Removed |
||
DreadPoor_Damasteel-8B-LINEAR_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Damasteel-8B-LINEAR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Damasteel-8B-LINEAR</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Damasteel-8B-LINEAR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Damasteel-8B-LINEAR | cfc389c15e614b14f1d8d16740dcc183047b435a | 28.964891 | 0 | 8 | false | false | false | true | 0.674659 | 0.738442 | 73.844178 | 0.538814 | 34.106138 | 0.166163 | 16.616314 | 0.298658 | 6.487696 | 0.42125 | 11.85625 | 0.377909 | 30.878768 | false | false | 2024-11-01 | 0 | Removed |
||
DreadPoor_Dearly_Beloved-8B-TIES_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Dearly_Beloved-8B-TIES" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Dearly_Beloved-8B-TIES</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Dearly_Beloved-8B-TIES-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Dearly_Beloved-8B-TIES | af6515ee730d6aa17d77687fe2c06c57fa9533fb | 26.034998 | 0 | 8 | false | false | false | true | 0.715241 | 0.826669 | 82.666879 | 0.404983 | 16.671813 | 0.196375 | 19.637462 | 0.298658 | 6.487696 | 0.417469 | 10.45026 | 0.282663 | 20.295878 | false | false | 2024-11-22 | 0 | Removed |
||
DreadPoor_Emu_Eggs-9B-Model_Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Emu_Eggs-9B-Model_Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Emu_Eggs-9B-Model_Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Emu_Eggs-9B-Model_Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Emu_Eggs-9B-Model_Stock | 3fb1b2da72f3618f6943aedfd1600df27886792a | 29.611715 | apache-2.0 | 2 | 9 | true | false | false | true | 3.08835 | 0.760698 | 76.069828 | 0.605166 | 42.783674 | 0.02568 | 2.567976 | 0.333054 | 11.073826 | 0.407083 | 9.31875 | 0.422706 | 35.856235 | true | false | 2024-10-18 | 2024-10-18 | 0 | DreadPoor/Emu_Eggs-9B-Model_Stock |
DreadPoor_Eunoia_Vespera-8B-LINEAR_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Eunoia_Vespera-8B-LINEAR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Eunoia_Vespera-8B-LINEAR</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Eunoia_Vespera-8B-LINEAR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Eunoia_Vespera-8B-LINEAR | c674956327af664735cf39b20c7a8276dfa579f9 | 28.931156 | apache-2.0 | 2 | 8 | true | false | false | true | 0.81326 | 0.723529 | 72.352912 | 0.539931 | 34.216103 | 0.152568 | 15.256798 | 0.307047 | 7.606264 | 0.41849 | 12.611198 | 0.383893 | 31.543661 | true | false | 2024-09-22 | 2024-09-22 | 1 | DreadPoor/Eunoia_Vespera-8B-LINEAR (Merge) |
DreadPoor_Heart_Stolen-8B-Model_Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Heart_Stolen-8B-Model_Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Heart_Stolen-8B-Model_Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Heart_Stolen-8B-Model_Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Heart_Stolen-8B-Model_Stock | 6d77987af7115c7455ddb072c48316815b018999 | 29.24739 | cc-by-nc-4.0 | 5 | 8 | true | false | false | true | 0.749301 | 0.724453 | 72.445334 | 0.539544 | 34.444822 | 0.162387 | 16.238671 | 0.317114 | 8.948546 | 0.416229 | 12.361979 | 0.379405 | 31.044991 | true | false | 2024-09-09 | 2024-09-10 | 1 | DreadPoor/Heart_Stolen-8B-Model_Stock (Merge) |
DreadPoor_Heart_Stolen-ALT-8B-Model_Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Heart_Stolen-ALT-8B-Model_Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Heart_Stolen-ALT-8B-Model_Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Heart_Stolen-ALT-8B-Model_Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Heart_Stolen-ALT-8B-Model_Stock | 03d1d70cb7eb5a743468b97c9c580028df487564 | 27.754545 | 0 | 8 | false | false | false | true | 0.735627 | 0.718358 | 71.83584 | 0.526338 | 32.354424 | 0.149547 | 14.954683 | 0.301174 | 6.823266 | 0.4055 | 9.754167 | 0.377244 | 30.804891 | false | false | 2024-09-11 | 0 | Removed |
||
DreadPoor_Irina-8B-model_stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Irina-8B-model_stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Irina-8B-model_stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Irina-8B-model_stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Irina-8B-model_stock | b282e3ab449d71a31f48b8c13eb43a4435968728 | 25.32468 | 0 | 8 | false | false | false | true | 0.745586 | 0.67994 | 67.994034 | 0.523664 | 32.08833 | 0.100453 | 10.045317 | 0.284396 | 4.58613 | 0.400292 | 8.636458 | 0.35738 | 28.597813 | false | false | 2024-08-30 | 0 | Removed |
||
DreadPoor_L3.1-BaeZel-8B-Della_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/L3.1-BaeZel-8B-Della" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/L3.1-BaeZel-8B-Della</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__L3.1-BaeZel-8B-Della-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/L3.1-BaeZel-8B-Della | ec61b6f5355a7f3975d80f1afac69e0407e612e5 | 26.167555 | 0 | 8 | false | false | false | true | 0.658995 | 0.518024 | 51.80244 | 0.544845 | 35.157455 | 0.169184 | 16.918429 | 0.319631 | 9.284116 | 0.419979 | 11.597396 | 0.390209 | 32.245493 | false | false | 2024-11-15 | 0 | Removed |
||
DreadPoor_Matryoshka-8B-LINEAR_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Matryoshka-8B-LINEAR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Matryoshka-8B-LINEAR</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Matryoshka-8B-LINEAR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Matryoshka-8B-LINEAR | 20d260e6d881fcd3b4f76071797675d095ba8e98 | 29.825026 | 0 | 8 | false | false | false | true | 0.662314 | 0.726252 | 72.62519 | 0.544428 | 35.110912 | 0.175227 | 17.522659 | 0.32047 | 9.395973 | 0.42525 | 12.45625 | 0.386553 | 31.83917 | false | false | 2024-12-02 | 0 | Removed |
||
DreadPoor_Mercury_In_Retrograde-8b-Model-Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Mercury_In_Retrograde-8b-Model-Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Mercury_In_Retrograde-8b-Model-Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Mercury_In_Retrograde-8b-Model-Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Mercury_In_Retrograde-8b-Model-Stock | 6c761644a57ab267624987ec2211c4af7a51a16a | 29.236114 | 0 | 8 | false | false | false | true | 0.673547 | 0.729624 | 72.962406 | 0.539051 | 34.384865 | 0.163142 | 16.314199 | 0.316275 | 8.836689 | 0.419885 | 11.485677 | 0.382896 | 31.432846 | false | false | 2024-12-03 | 0 | Removed |
||
DreadPoor_ONeil-model_stock-8B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/ONeil-model_stock-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/ONeil-model_stock-8B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__ONeil-model_stock-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/ONeil-model_stock-8B | d4b84956211fd57b85122fe0c6f88b2cb9a9c86a | 26.935908 | cc-by-nc-4.0 | 2 | 8 | true | false | false | true | 0.764449 | 0.678566 | 67.85662 | 0.554834 | 36.412613 | 0.101208 | 10.120846 | 0.305369 | 7.38255 | 0.417344 | 10.967969 | 0.359874 | 28.874852 | true | false | 2024-07-06 | 2024-07-15 | 1 | DreadPoor/ONeil-model_stock-8B (Merge) |
DreadPoor_Promissum_Mane-8B-LINEAR_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Promissum_Mane-8B-LINEAR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Promissum_Mane-8B-LINEAR</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Promissum_Mane-8B-LINEAR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Promissum_Mane-8B-LINEAR | ff399e7004040e1807e8d08b4d0967206fc50afa | 29.049297 | 1 | 8 | false | false | false | true | 0.828647 | 0.715036 | 71.50361 | 0.545768 | 35.25319 | 0.152568 | 15.256798 | 0.30453 | 7.270694 | 0.420042 | 13.338542 | 0.385057 | 31.672946 | false | false | 2024-09-30 | 2024-09-30 | 1 | DreadPoor/Promissum_Mane-8B-LINEAR (Merge) |
|
DreadPoor_Promissum_Mane-8B-LINEAR-lorablated_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Promissum_Mane-8B-LINEAR-lorablated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Promissum_Mane-8B-LINEAR-lorablated</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Promissum_Mane-8B-LINEAR-lorablated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Promissum_Mane-8B-LINEAR-lorablated | 34c4a30b7462704810e35e033aa5ef33b075a97b | 28.810739 | 0 | 8 | false | false | false | true | 0.792342 | 0.715636 | 71.563562 | 0.543518 | 34.609107 | 0.152568 | 15.256798 | 0.303691 | 7.158837 | 0.419792 | 13.840625 | 0.37392 | 30.435505 | false | false | 2024-09-30 | 0 | Removed |
||
DreadPoor_Sellen-8B-model_stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Sellen-8B-model_stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Sellen-8B-model_stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Sellen-8B-model_stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Sellen-8B-model_stock | accde7145d81a428c782695ea61eebc608efd980 | 26.362467 | 0 | 8 | false | false | false | true | 0.807471 | 0.711289 | 71.128938 | 0.523168 | 31.360979 | 0.132175 | 13.217523 | 0.274329 | 3.243848 | 0.396042 | 10.671875 | 0.356965 | 28.55164 | false | false | 2024-08-27 | 0 | Removed |
||
DreadPoor_Sweetened_Condensed_Milk-8B-Model_Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Sweetened_Condensed_Milk-8B-Model_Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Sweetened_Condensed_Milk-8B-Model_Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Sweetened_Condensed_Milk-8B-Model_Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Sweetened_Condensed_Milk-8B-Model_Stock | 307d93aeb51160a8b0ce236b8abd13e04873fef1 | 29.523394 | 0 | 8 | false | false | false | true | 0.665669 | 0.741714 | 74.171421 | 0.540629 | 34.670888 | 0.185045 | 18.504532 | 0.302852 | 7.04698 | 0.410677 | 11.101302 | 0.384807 | 31.645242 | false | false | 2024-11-27 | 0 | Removed |
||
DreadPoor_Trinas_Nectar-8B-model_stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/Trinas_Nectar-8B-model_stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/Trinas_Nectar-8B-model_stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__Trinas_Nectar-8B-model_stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/Trinas_Nectar-8B-model_stock | cb46b8431872557904d83fc5aa1b90dabeb74acc | 27.535042 | cc-by-nc-4.0 | 3 | 8 | true | false | false | true | 0.866724 | 0.725927 | 72.592721 | 0.525612 | 31.975094 | 0.153323 | 15.332326 | 0.286074 | 4.809843 | 0.406771 | 11.413021 | 0.361785 | 29.087249 | true | false | 2024-08-16 | 2024-08-27 | 1 | DreadPoor/Trinas_Nectar-8B-model_stock (Merge) |
DreadPoor_WIP-Acacia-8B-Model_Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/WIP-Acacia-8B-Model_Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/WIP-Acacia-8B-Model_Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__WIP-Acacia-8B-Model_Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/WIP-Acacia-8B-Model_Stock | ae4f1a21b9de70ec75d02e9a84209ae6360a01e9 | 26.560174 | 0 | 8 | false | false | false | true | 0.662951 | 0.624636 | 62.463597 | 0.519467 | 31.162353 | 0.1571 | 15.70997 | 0.306208 | 7.494407 | 0.422583 | 12.122917 | 0.37367 | 30.407801 | false | false | 2024-11-28 | 0 | Removed |
||
DreadPoor_WIP_Damascus-8B-TIES_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/WIP_Damascus-8B-TIES" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/WIP_Damascus-8B-TIES</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__WIP_Damascus-8B-TIES-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/WIP_Damascus-8B-TIES | c7720a0b0a8d24e62bf71b0e955b1aca8e62f1cb | 24.731381 | 0 | 8 | false | false | false | true | 0.818112 | 0.477633 | 47.763268 | 0.541067 | 34.522306 | 0.151057 | 15.10574 | 0.307047 | 7.606264 | 0.411854 | 12.715104 | 0.37608 | 30.675606 | false | false | 2024-10-29 | 0 | Removed |
||
DreadPoor_felix_dies-mistral-7B-model_stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/felix_dies-mistral-7B-model_stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/felix_dies-mistral-7B-model_stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__felix_dies-mistral-7B-model_stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/felix_dies-mistral-7B-model_stock | bb317aa7565625327e18c5158aebd4710aa1d925 | 18.101828 | 0 | 7 | false | false | false | false | 0.661572 | 0.300779 | 30.07786 | 0.490092 | 28.890798 | 0.05136 | 5.135952 | 0.291946 | 5.592841 | 0.451823 | 15.477865 | 0.310921 | 23.435653 | false | false | 2024-09-30 | 0 | Removed |
||
DreadPoor_remember_to_breathe-8b-Model-Stock_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./DreadPoor/remember_to_breathe-8b-Model-Stock" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">DreadPoor/remember_to_breathe-8b-Model-Stock</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/DreadPoor__remember_to_breathe-8b-Model-Stock-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | DreadPoor/remember_to_breathe-8b-Model-Stock | fa88f1b06cf9ca7bd0d859c6a4b2240485363ae0 | 28.168407 | 0 | 8 | false | false | false | true | 0.66354 | 0.710415 | 71.041503 | 0.541165 | 34.678991 | 0.143505 | 14.350453 | 0.301174 | 6.823266 | 0.414458 | 11.440625 | 0.37608 | 30.675606 | false | false | 2024-12-06 | 0 | Removed |
||
EVA-UNIT-01_EVA-Qwen2.5-72B-v0.2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co./EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EVA-UNIT-01__EVA-Qwen2.5-72B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EVA-UNIT-01/EVA-Qwen2.5-72B-v0.2 | 2590214b30391392b9a84e7cbe40fff3a92c6814 | 43.541838 | other | 9 | 72 | true | false | false | true | 22.955098 | 0.687884 | 68.78837 | 0.708801 | 59.066733 | 0.390483 | 39.048338 | 0.408557 | 21.14094 | 0.471979 | 19.730729 | 0.581283 | 53.475916 | false | false | 2024-11-21 | 2024-11-27 | 1 | Qwen/Qwen2.5-72B |
EleutherAI_gpt-j-6b_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | GPTJForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/gpt-j-6b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/gpt-j-6b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__gpt-j-6b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/gpt-j-6b | 47e169305d2e8376be1d31e765533382721b2cc1 | 6.557824 | apache-2.0 | 1,450 | 6 | true | false | false | false | 0.767432 | 0.252219 | 25.221856 | 0.319104 | 4.912818 | 0.01284 | 1.283988 | 0.245805 | 0 | 0.36575 | 5.252083 | 0.124086 | 2.676197 | false | true | 2022-03-02 | 2024-08-19 | 0 | EleutherAI/gpt-j-6b |
EleutherAI_gpt-neo-1.3B_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | GPTNeoForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/gpt-neo-1.3B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/gpt-neo-1.3B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__gpt-neo-1.3B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/gpt-neo-1.3B | dbe59a7f4a88d01d1ba9798d78dbe3fe038792c8 | 5.340738 | mit | 269 | 1 | true | false | false | false | 0.359424 | 0.207905 | 20.790503 | 0.303923 | 3.024569 | 0.007553 | 0.755287 | 0.255872 | 0.782998 | 0.381656 | 4.873698 | 0.116356 | 1.817376 | false | true | 2022-03-02 | 2024-06-12 | 0 | EleutherAI/gpt-neo-1.3B |
EleutherAI_gpt-neo-125m_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | GPTNeoForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/gpt-neo-125m" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/gpt-neo-125m</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__gpt-neo-125m-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/gpt-neo-125m | 21def0189f5705e2521767faed922f1f15e7d7db | 4.382146 | mit | 184 | 0 | true | false | false | false | 0.202902 | 0.190544 | 19.054442 | 0.311516 | 3.436739 | 0.004532 | 0.453172 | 0.253356 | 0.447427 | 0.359333 | 2.616667 | 0.10256 | 0.284427 | false | true | 2022-03-02 | 2024-08-10 | 0 | EleutherAI/gpt-neo-125m |
EleutherAI_gpt-neo-2.7B_bfloat16 | bfloat16 | 🟢 pretrained | 🟢 | Original | GPTNeoForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/gpt-neo-2.7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/gpt-neo-2.7B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__gpt-neo-2.7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/gpt-neo-2.7B | e24fa291132763e59f4a5422741b424fb5d59056 | 6.355519 | mit | 444 | 2 | true | false | false | false | 0.508381 | 0.258963 | 25.896289 | 0.313952 | 4.178603 | 0.006042 | 0.60423 | 0.26594 | 2.12528 | 0.355365 | 3.520573 | 0.116273 | 1.808141 | false | true | 2022-03-02 | 2024-06-12 | 0 | EleutherAI/gpt-neo-2.7B |
EleutherAI_gpt-neox-20b_float16 | float16 | 🟢 pretrained | 🟢 | Original | GPTNeoXForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/gpt-neox-20b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/gpt-neox-20b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__gpt-neox-20b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/gpt-neox-20b | c292233c833e336628618a88a648727eb3dff0a7 | 6.003229 | apache-2.0 | 543 | 20 | true | false | false | false | 3.146736 | 0.258688 | 25.868806 | 0.316504 | 4.929114 | 0.006798 | 0.679758 | 0.243289 | 0 | 0.364667 | 2.816667 | 0.115525 | 1.72503 | false | true | 2022-04-07 | 2024-06-09 | 0 | EleutherAI/gpt-neox-20b |
EleutherAI_pythia-12b_float16 | float16 | 🟢 pretrained | 🟢 | Original | GPTNeoXForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/pythia-12b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/pythia-12b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__pythia-12b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/pythia-12b | 35c9d7f32fbb108fb8b5bdd574eb03369d1eed49 | 5.93396 | apache-2.0 | 131 | 12 | true | false | false | false | 1.118007 | 0.247148 | 24.714757 | 0.317965 | 4.987531 | 0.009063 | 0.906344 | 0.246644 | 0 | 0.364698 | 3.78724 | 0.110871 | 1.20789 | false | true | 2023-02-28 | 2024-06-12 | 0 | EleutherAI/pythia-12b |
EleutherAI_pythia-160m_float16 | float16 | 🟢 pretrained | 🟢 | Original | GPTNeoXForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/pythia-160m" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/pythia-160m</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__pythia-160m-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/pythia-160m | 50f5173d932e8e61f858120bcb800b97af589f46 | 5.617102 | apache-2.0 | 26 | 0 | true | false | false | false | 0.235339 | 0.181552 | 18.155162 | 0.297044 | 2.198832 | 0.002266 | 0.226586 | 0.258389 | 1.118568 | 0.417938 | 10.675521 | 0.111951 | 1.32794 | false | true | 2023-02-08 | 2024-06-09 | 0 | EleutherAI/pythia-160m |
EleutherAI_pythia-2.8b_float16 | float16 | 🟢 pretrained | 🟢 | Original | GPTNeoXForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/pythia-2.8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/pythia-2.8b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__pythia-2.8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/pythia-2.8b | 2a259cdd96a4beb1cdf467512e3904197345f6a9 | 5.454241 | apache-2.0 | 29 | 2 | true | false | false | false | 0.753902 | 0.217322 | 21.732226 | 0.322409 | 5.077786 | 0.007553 | 0.755287 | 0.25 | 0 | 0.348573 | 3.638281 | 0.113697 | 1.521868 | false | true | 2023-02-13 | 2024-06-12 | 0 | EleutherAI/pythia-2.8b |
EleutherAI_pythia-410m_float16 | float16 | 🟢 pretrained | 🟢 | Original | GPTNeoXForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/pythia-410m" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/pythia-410m</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__pythia-410m-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/pythia-410m | 9879c9b5f8bea9051dcb0e68dff21493d67e9d4f | 5.113779 | apache-2.0 | 22 | 0 | true | false | false | false | 0.377082 | 0.219545 | 21.954525 | 0.302813 | 2.715428 | 0.003021 | 0.302115 | 0.259228 | 1.230425 | 0.357813 | 3.059896 | 0.112783 | 1.420287 | false | true | 2023-02-13 | 2024-06-09 | 0 | EleutherAI/pythia-410m |
EleutherAI_pythia-6.9b_float16 | float16 | 🟢 pretrained | 🟢 | Original | GPTNeoXForCausalLM | <a target="_blank" href="https://huggingface.co./EleutherAI/pythia-6.9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EleutherAI/pythia-6.9b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EleutherAI__pythia-6.9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EleutherAI/pythia-6.9b | f271943e880e60c0c715fd10e4dc74ec4e31eb44 | 5.865842 | apache-2.0 | 48 | 6 | true | false | false | false | 0.868867 | 0.228114 | 22.811363 | 0.323229 | 5.881632 | 0.008308 | 0.830816 | 0.251678 | 0.223714 | 0.359052 | 3.814844 | 0.114694 | 1.632683 | false | true | 2023-02-14 | 2024-06-12 | 0 | EleutherAI/pythia-6.9b |
Enno-Ai_EnnoAi-Pro-French-Llama-3-8B-v0.4_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Enno-Ai/EnnoAi-Pro-French-Llama-3-8B-v0.4" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Enno-Ai/EnnoAi-Pro-French-Llama-3-8B-v0.4</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-French-Llama-3-8B-v0.4-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Enno-Ai/EnnoAi-Pro-French-Llama-3-8B-v0.4 | 328722ae96e3a112ec900dbe77d410788a526c5c | 15.180945 | creativeml-openrail-m | 0 | 8 | true | false | false | true | 1.009128 | 0.418881 | 41.888079 | 0.407495 | 16.875928 | 0.006042 | 0.60423 | 0.270973 | 2.796421 | 0.417 | 10.758333 | 0.263464 | 18.162677 | false | false | 2024-06-27 | 2024-06-30 | 0 | Enno-Ai/EnnoAi-Pro-French-Llama-3-8B-v0.4 |
Enno-Ai_EnnoAi-Pro-Llama-3-8B_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Enno-Ai/EnnoAi-Pro-Llama-3-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Enno-Ai/EnnoAi-Pro-Llama-3-8B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Enno-Ai/EnnoAi-Pro-Llama-3-8B | 6a5d745bdd304753244fe601e2a958d37d13cd71 | 12.174667 | creativeml-openrail-m | 0 | 8 | true | false | false | true | 1.184337 | 0.319538 | 31.953772 | 0.415158 | 17.507545 | 0.001511 | 0.151057 | 0.261745 | 1.565996 | 0.407052 | 9.08151 | 0.215093 | 12.788121 | false | false | 2024-07-01 | 2024-07-08 | 0 | Enno-Ai/EnnoAi-Pro-Llama-3-8B |
Enno-Ai_EnnoAi-Pro-Llama-3-8B-v0.3_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3-8B-v0.3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3 | cf29b8b484a909132e3a1f85ce891d28347c0d13 | 17.524058 | creativeml-openrail-m | 0 | 8 | true | false | false | true | 1.470836 | 0.508257 | 50.825698 | 0.410058 | 16.668386 | 0.012085 | 1.208459 | 0.265101 | 2.013423 | 0.423573 | 12.313281 | 0.299036 | 22.1151 | false | false | 2024-06-26 | 2024-06-26 | 0 | Enno-Ai/EnnoAi-Pro-Llama-3-8B-v0.3 |
Enno-Ai_EnnoAi-Pro-Llama-3.1-8B-v0.9_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Enno-Ai/EnnoAi-Pro-Llama-3.1-8B-v0.9" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Enno-Ai/EnnoAi-Pro-Llama-3.1-8B-v0.9</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Enno-Ai__EnnoAi-Pro-Llama-3.1-8B-v0.9-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Enno-Ai/EnnoAi-Pro-Llama-3.1-8B-v0.9 | c740871122fd471a1a225cf2b4368e333752d74c | 14.945694 | apache-2.0 | 0 | 8 | true | false | false | true | 0.932571 | 0.468915 | 46.89147 | 0.416027 | 17.498296 | 0 | 0 | 0.26594 | 2.12528 | 0.383177 | 5.430469 | 0.259558 | 17.72865 | false | false | 2024-08-22 | 2024-09-06 | 0 | Enno-Ai/EnnoAi-Pro-Llama-3.1-8B-v0.9 |
EnnoAi_EnnoAi-Pro-Llama-3.1-8B-v1.0_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EnnoAi/EnnoAi-Pro-Llama-3.1-8B-v1.0" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EnnoAi/EnnoAi-Pro-Llama-3.1-8B-v1.0</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EnnoAi__EnnoAi-Pro-Llama-3.1-8B-v1.0-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EnnoAi/EnnoAi-Pro-Llama-3.1-8B-v1.0 | c740871122fd471a1a225cf2b4368e333752d74c | 14.97109 | apache-2.0 | 0 | 8 | true | false | false | true | 0.945642 | 0.470438 | 47.043844 | 0.416027 | 17.498296 | 0 | 0 | 0.26594 | 2.12528 | 0.383177 | 5.430469 | 0.259558 | 17.72865 | false | false | 2024-08-22 | 2024-09-06 | 0 | EnnoAi/EnnoAi-Pro-Llama-3.1-8B-v1.0 |
Epiculous_Azure_Dusk-v0.2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./Epiculous/Azure_Dusk-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Epiculous/Azure_Dusk-v0.2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Epiculous__Azure_Dusk-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Epiculous/Azure_Dusk-v0.2 | ebddf1b2efbe7f9cae066d263b0991ded89c88e8 | 14.050827 | apache-2.0 | 7 | 12 | true | false | false | true | 1.991411 | 0.346716 | 34.67156 | 0.411972 | 17.396414 | 0.018127 | 1.812689 | 0.260906 | 1.454139 | 0.383458 | 6.365625 | 0.303441 | 22.604536 | false | false | 2024-09-09 | 2024-09-14 | 0 | Epiculous/Azure_Dusk-v0.2 |
Epiculous_Crimson_Dawn-v0.2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./Epiculous/Crimson_Dawn-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Epiculous/Crimson_Dawn-v0.2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Epiculous__Crimson_Dawn-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Epiculous/Crimson_Dawn-v0.2 | 4cceb1e25026afef241ad5325097e88eccd8f37a | 14.884541 | apache-2.0 | 10 | 12 | true | false | false | true | 3.492384 | 0.310345 | 31.034544 | 0.448238 | 21.688249 | 0.030967 | 3.096677 | 0.276007 | 3.467562 | 0.415177 | 10.897135 | 0.272108 | 19.123079 | false | false | 2024-09-02 | 2024-09-05 | 0 | Epiculous/Crimson_Dawn-v0.2 |
Epiculous_NovaSpark_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Epiculous/NovaSpark" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Epiculous/NovaSpark</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Epiculous__NovaSpark-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Epiculous/NovaSpark | a46340895859e470c3e69661f0b894677cf4c5cb | 25.228562 | apache-2.0 | 6 | 8 | true | false | false | true | 0.818185 | 0.640847 | 64.08474 | 0.506396 | 29.526911 | 0.150302 | 15.030211 | 0.297819 | 6.375839 | 0.388198 | 6.92474 | 0.36486 | 29.42893 | false | false | 2024-10-13 | 2024-10-20 | 1 | Epiculous/NovaSpark (Merge) |
Epiculous_Violet_Twilight-v0.2_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./Epiculous/Violet_Twilight-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Epiculous/Violet_Twilight-v0.2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Epiculous__Violet_Twilight-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Epiculous/Violet_Twilight-v0.2 | 30c8bad3c1f565150afbf2fc90cacf4f45d096f6 | 18.552773 | apache-2.0 | 17 | 12 | true | false | false | true | 1.770436 | 0.453178 | 45.317757 | 0.461455 | 23.940537 | 0.028701 | 2.870091 | 0.26594 | 2.12528 | 0.429938 | 13.608854 | 0.311087 | 23.454122 | true | false | 2024-09-12 | 2024-09-16 | 0 | Epiculous/Violet_Twilight-v0.2 |
EpistemeAI_Alpaca-Llama3.1-8B_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Alpaca-Llama3.1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Alpaca-Llama3.1-8B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Alpaca-Llama3.1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Alpaca-Llama3.1-8B | 3152dfa17322dff7c6af6dbf3daceaf5db51e230 | 13.922106 | apache-2.0 | 0 | 8 | true | false | false | false | 0.920853 | 0.159869 | 15.986915 | 0.475526 | 25.935227 | 0.046828 | 4.682779 | 0.290268 | 5.369128 | 0.34026 | 6.599219 | 0.324634 | 24.959368 | false | false | 2024-09-11 | 2024-08-13 | 2 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI_Athena-gemma-2-2b-it_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Athena-gemma-2-2b-it" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Athena-gemma-2-2b-it</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Athena-gemma-2-2b-it-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Athena-gemma-2-2b-it | 661c1dc6a1a096222e33416e099bd02b7b970405 | 14.294329 | apache-2.0 | 2 | 2 | true | false | false | false | 2.036798 | 0.313417 | 31.341729 | 0.426423 | 19.417818 | 0.033988 | 3.398792 | 0.268456 | 2.46085 | 0.435052 | 13.348177 | 0.242188 | 15.798611 | false | false | 2024-08-29 | 2024-09-06 | 4 | google/gemma-2-9b |
EpistemeAI_Athena-gemma-2-2b-it-Philos_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Gemma2ForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Athena-gemma-2-2b-it-Philos" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Athena-gemma-2-2b-it-Philos</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Athena-gemma-2-2b-it-Philos-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Athena-gemma-2-2b-it-Philos | dea2b35d496bd32ed3c88d42ff3022654153f2e1 | 15.122657 | apache-2.0 | 0 | 2 | true | false | false | true | 1.128593 | 0.462095 | 46.209502 | 0.379478 | 13.212088 | 0.004532 | 0.453172 | 0.28104 | 4.138702 | 0.431365 | 12.853906 | 0.224817 | 13.868573 | false | false | 2024-09-05 | 2024-09-05 | 1 | unsloth/gemma-2-2b-it-bnb-4bit |
EpistemeAI_Athene-codegemma-2-7b-it-alpaca-v1.3_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | GemmaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Athene-codegemma-2-7b-it-alpaca-v1.3" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Athene-codegemma-2-7b-it-alpaca-v1.3</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Athene-codegemma-2-7b-it-alpaca-v1.3-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Athene-codegemma-2-7b-it-alpaca-v1.3 | 9c26e1242a11178b53937bc0e9a744ef6141e05a | 17.314022 | apache-2.0 | 0 | 7 | true | false | false | false | 0.971978 | 0.402994 | 40.299406 | 0.433192 | 20.873795 | 0.061934 | 6.193353 | 0.280201 | 4.026846 | 0.450302 | 14.854427 | 0.258727 | 17.636303 | false | false | 2024-09-06 | 2024-09-06 | 2 | Removed |
EpistemeAI_FineLlama3.1-8B-Instruct_4bit | 4bit | 🔶 fine-tuned on domain-specific datasets | 🔶 | Adapter | ? | <a target="_blank" href="https://huggingface.co./EpistemeAI/FineLlama3.1-8B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/FineLlama3.1-8B-Instruct</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__FineLlama3.1-8B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/FineLlama3.1-8B-Instruct | a8b0fc584b10e0110e04f9d21c7f10d24391c1d5 | 11.100787 | 0 | 14 | false | false | false | false | 2.354961 | 0.08001 | 8.000993 | 0.455736 | 23.506619 | 0.026435 | 2.643505 | 0.280201 | 4.026846 | 0.348167 | 4.954167 | 0.311253 | 23.472592 | false | false | 2024-08-10 | 0 | Removed |
||
EpistemeAI_Fireball-12B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-12B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-12B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-12B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-12B | e2ed12c3244f2502321fb20e76dfc72ad7817d6e | 15.509355 | apache-2.0 | 1 | 12 | true | false | false | false | 1.618521 | 0.18335 | 18.335018 | 0.511089 | 30.666712 | 0.039275 | 3.927492 | 0.261745 | 1.565996 | 0.423635 | 12.521094 | 0.334358 | 26.03982 | false | false | 2024-08-20 | 2024-08-21 | 2 | Removed |
EpistemeAI_Fireball-12B-v1.13a-philosophers_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-12B-v1.13a-philosophers" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-12B-v1.13a-philosophers</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-12B-v1.13a-philosophers-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-12B-v1.13a-philosophers | 7fa824d4a40abca3f8c75d432ea151dc0d1d67d6 | 14.440865 | apache-2.0 | 2 | 12 | true | false | false | false | 1.662663 | 0.087553 | 8.755325 | 0.51027 | 30.336233 | 0.044562 | 4.456193 | 0.301174 | 6.823266 | 0.408073 | 9.975781 | 0.336686 | 26.298389 | false | false | 2024-08-28 | 2024-09-03 | 1 | Removed |
EpistemeAI_Fireball-Alpaca-Llama-3.1-8B-Philos-DPO-200_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Alpaca-Llama-3.1-8B-Philos-DPO-200" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Alpaca-Llama-3.1-8B-Philos-DPO-200</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Alpaca-Llama-3.1-8B-Philos-DPO-200-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Alpaca-Llama-3.1-8B-Philos-DPO-200 | 27d67626304954db71f21fec9e7fc516421274ec | 21.066974 | apache-2.0 | 0 | 8 | true | false | false | false | 0.922381 | 0.457724 | 45.772439 | 0.48384 | 26.377774 | 0.119335 | 11.933535 | 0.300336 | 6.711409 | 0.394458 | 6.907292 | 0.358295 | 28.699394 | false | false | 2024-09-16 | 2024-09-16 | 4 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI_Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta | 2851384717556dd6ac14c00ed87aac1f267eb263 | 25.179287 | apache-2.0 | 0 | 8 | true | false | false | true | 0.885645 | 0.727401 | 72.740107 | 0.486489 | 26.897964 | 0.148792 | 14.879154 | 0.280201 | 4.026846 | 0.361938 | 4.275521 | 0.354305 | 28.256132 | false | false | 2024-09-12 | 2024-09-14 | 5 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI_Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R2_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R2 | b19336101aa5f4807d1574f4c11eebc1c1a1c34e | 22.537889 | apache-2.0 | 0 | 8 | true | false | false | false | 0.811743 | 0.467316 | 46.731561 | 0.493203 | 28.247009 | 0.123112 | 12.311178 | 0.286074 | 4.809843 | 0.462365 | 16.995573 | 0.335189 | 26.132166 | false | false | 2024-09-14 | 2024-09-14 | 3 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-0.001-128K-auto_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-0.001-128K-auto" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-0.001-128K-auto</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-0.001-128K-auto-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-0.001-128K-auto | 19b23c434b6c4524e2146926cdbf4f0e927ae3ab | 21.57995 | apache-2.0 | 0 | 8 | true | false | false | false | 0.694994 | 0.443186 | 44.31863 | 0.482364 | 26.832967 | 0.133686 | 13.36858 | 0.312081 | 8.277405 | 0.406646 | 8.730729 | 0.351563 | 27.951389 | false | false | 2024-11-14 | 2024-11-15 | 2 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K | b4a88fb5fb27fc5d8a503303cdb7aaeff373fd92 | 20.627168 | apache-2.0 | 3 | 8 | true | false | false | false | 0.814786 | 0.445734 | 44.573399 | 0.489732 | 28.025161 | 0.120846 | 12.084592 | 0.294463 | 5.928412 | 0.376229 | 4.895312 | 0.354305 | 28.256132 | false | false | 2024-09-26 | 2024-10-05 | 1 | Removed |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code | 8e8f1569a8a01ed3d6588f2669c730d4993355b5 | 23.89695 | apache-2.0 | 2 | 8 | true | false | false | false | 0.854318 | 0.597533 | 59.753343 | 0.490419 | 28.171888 | 0.13142 | 13.141994 | 0.302013 | 6.935123 | 0.401031 | 8.46224 | 0.342254 | 26.91711 | false | false | 2024-10-04 | 2024-10-05 | 2 | Removed |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds | 8b73dd02349f0544c48c581cc73ada5cac6ff946 | 22.993108 | llama3.1 | 2 | 8 | true | false | false | true | 1.716734 | 0.669099 | 66.90991 | 0.466807 | 24.462654 | 0.124622 | 12.462236 | 0.272651 | 3.020134 | 0.341781 | 4.55599 | 0.33893 | 26.547725 | false | false | 2024-10-14 | 2024-10-15 | 4 | Removed |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto | f18598c62a783bcc0d436a35df0c8a335e8ee5d7 | 23.749941 | apache-2.0 | 6 | 8 | true | false | false | true | 2.285306 | 0.730498 | 73.049841 | 0.464925 | 24.586737 | 0.139728 | 13.97281 | 0.26594 | 2.12528 | 0.320885 | 1.210677 | 0.347989 | 27.5543 | false | false | 2024-10-21 | 2024-10-29 | 1 | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto (Merge) |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto | 055e87600d18e58594a8d193f45c0ee9a90e1780 | 23.488818 | apache-2.0 | 6 | 8 | true | false | false | true | 0.672068 | 0.720707 | 72.070661 | 0.461009 | 23.544253 | 0.123112 | 12.311178 | 0.270134 | 2.684564 | 0.34324 | 4.171615 | 0.335356 | 26.150635 | false | false | 2024-10-21 | 2024-11-27 | 1 | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto (Merge) |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-COT_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-COT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-COT</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-COT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-COT | bb90c19dc7c4a509e7bd73f4620dca818b58be25 | 20.832251 | apache-2.0 | 0 | 8 | true | false | false | false | 0.839037 | 0.457824 | 45.782413 | 0.476052 | 25.820865 | 0.136707 | 13.670695 | 0.293624 | 5.816555 | 0.388135 | 6.45026 | 0.347074 | 27.452719 | false | false | 2024-10-11 | 2024-10-11 | 3 | Removed |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-ds-auto_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-ds-auto" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-ds-auto</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-ds-auto-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-ds-auto | db5ddb161ed26bc16baa814e31892dbe2f22b7a0 | 23.760965 | apache-2.0 | 0 | 8 | true | false | false | true | 0.745131 | 0.720482 | 72.048166 | 0.48178 | 26.45206 | 0.136707 | 13.670695 | 0.248322 | 0 | 0.33 | 2.083333 | 0.354804 | 28.31154 | false | false | 2024-11-14 | 2024-11-14 | 1 | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.004-128K-code-ds-auto (Merge) |
EpistemeAI_Fireball-Meta-Llama-3.1-8B-Instruct-Math_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Math" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Math</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.1-8B-Instruct-Math-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Math | 677c97b4f92bfc330d4fae628e9a1df1ef606dcc | 20.545341 | apache-2.0 | 0 | 8 | true | false | false | false | 0.910272 | 0.462296 | 46.22956 | 0.498295 | 28.959344 | 0.107251 | 10.725076 | 0.291107 | 5.480984 | 0.364073 | 5.975781 | 0.333112 | 25.9013 | false | false | 2024-09-23 | 2024-09-23 | 3 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI_Fireball-Meta-Llama-3.2-8B-Instruct-agent-003-128k-code-DPO_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Meta-Llama-3.2-8B-Instruct-agent-003-128k-code-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Meta-Llama-3.2-8B-Instruct-agent-003-128k-code-DPO</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Meta-Llama-3.2-8B-Instruct-agent-003-128k-code-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Meta-Llama-3.2-8B-Instruct-agent-003-128k-code-DPO | b3c0fce7daa359cd8ed5be6595dd1a76ca2cfea2 | 21.205445 | apache-2.0 | 1 | 8 | true | false | false | false | 0.833576 | 0.461097 | 46.109656 | 0.480101 | 26.317878 | 0.120091 | 12.009063 | 0.300336 | 6.711409 | 0.399823 | 8.077865 | 0.352061 | 28.006797 | false | false | 2024-10-08 | 2024-10-09 | 3 | Removed |
EpistemeAI_Fireball-Mistral-Nemo-Base-2407-v1-DPO2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Fireball-Mistral-Nemo-Base-2407-v1-DPO2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Fireball-Mistral-Nemo-Base-2407-v1-DPO2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Fireball-Mistral-Nemo-Base-2407-v1-DPO2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Fireball-Mistral-Nemo-Base-2407-v1-DPO2 | 2cf732fbffefdf37341b946edd7995f14d3f9487 | 15.2764 | apache-2.0 | 0 | 12 | true | false | false | false | 1.771269 | 0.186073 | 18.607295 | 0.496777 | 28.567825 | 0.032477 | 3.247734 | 0.291946 | 5.592841 | 0.40401 | 9.501302 | 0.335273 | 26.141401 | false | false | 2024-08-19 | 2024-08-19 | 1 | Removed |
EpistemeAI_Llama-3.2-3B-Agent007-Coder_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Llama-3.2-3B-Agent007-Coder" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Llama-3.2-3B-Agent007-Coder</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Llama-3.2-3B-Agent007-Coder-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Llama-3.2-3B-Agent007-Coder | 7ff4e77796b6d308e96d0150e1a01081c0b82e01 | 18.901974 | apache-2.0 | 0 | 3 | true | false | false | false | 0.710816 | 0.539956 | 53.995621 | 0.430376 | 19.025809 | 0.110272 | 11.02719 | 0.25755 | 1.006711 | 0.366802 | 7.783594 | 0.285156 | 20.572917 | false | false | 2024-10-08 | 2024-10-08 | 2 | meta-llama/Llama-3.2-3B-Instruct |
EpistemeAI_Mistral-Nemo-Instruct-12B-Philosophy-Math_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI/Mistral-Nemo-Instruct-12B-Philosophy-Math" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI/Mistral-Nemo-Instruct-12B-Philosophy-Math</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI__Mistral-Nemo-Instruct-12B-Philosophy-Math-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI/Mistral-Nemo-Instruct-12B-Philosophy-Math | 1ac4205f8da109326b4a5cf173e5491a20087d76 | 16.566232 | apache-2.0 | 0 | 12 | true | false | false | false | 1.363607 | 0.069468 | 6.94679 | 0.536493 | 33.835811 | 0.093656 | 9.365559 | 0.331376 | 10.850112 | 0.429219 | 12.885677 | 0.329621 | 25.513446 | false | false | 2024-09-15 | 2024-09-26 | 1 | unsloth/Mistral-Nemo-Instruct-2407-bnb-4bit |
EpistemeAI2_Athene-codegemma-2-7b-it-alpaca-v1.2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | GemmaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Athene-codegemma-2-7b-it-alpaca-v1.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Athene-codegemma-2-7b-it-alpaca-v1.2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Athene-codegemma-2-7b-it-alpaca-v1.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Athene-codegemma-2-7b-it-alpaca-v1.2 | 21b31062334a316b50680e8c3a141a72e4c30b61 | 15.693215 | apache-2.0 | 0 | 7 | true | false | false | false | 0.969635 | 0.435118 | 43.511771 | 0.417542 | 18.97137 | 0.040785 | 4.07855 | 0.270973 | 2.796421 | 0.416969 | 10.38776 | 0.229721 | 14.413416 | false | false | 2024-08-26 | 2024-08-26 | 2 | Removed |
EpistemeAI2_Fireball-12B-v1.2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-12B-v1.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-12B-v1.2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-12B-v1.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-12B-v1.2 | 57af42edf8232189ee99e9a21e33a0c306e3f561 | 15.162522 | apache-2.0 | 1 | 12 | true | false | false | false | 1.872565 | 0.135539 | 13.553926 | 0.501858 | 29.776014 | 0.039275 | 3.927492 | 0.298658 | 6.487696 | 0.417313 | 11.264062 | 0.333693 | 25.965943 | false | false | 2024-08-27 | 2024-08-28 | 1 | Removed |
EpistemeAI2_Fireball-Alpaca-Llama3.1-8B-Philos_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1-8B-Philos" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1-8B-Philos</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1-8B-Philos-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1-8B-Philos | 3dcca4cf9bdd9003c8dc91f5c78cefef1d4ae0d7 | 22.539085 | apache-2.0 | 1 | 8 | true | false | false | false | 0.848332 | 0.49864 | 49.864027 | 0.497758 | 29.259226 | 0.117825 | 11.782477 | 0.292785 | 5.704698 | 0.427667 | 11.891667 | 0.340592 | 26.732417 | false | false | 2024-08-29 | 2024-08-29 | 3 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Alpaca-Llama3.1.01-8B-Philos_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1.01-8B-Philos" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1.01-8B-Philos</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1.01-8B-Philos-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1.01-8B-Philos | f97293ed5cec7fb9482b16600259967c6c923e4b | 21.567144 | apache-2.0 | 0 | 8 | true | false | false | false | 0.870572 | 0.421179 | 42.117914 | 0.495611 | 28.628475 | 0.135952 | 13.595166 | 0.288591 | 5.145414 | 0.437062 | 13.432813 | 0.338348 | 26.483082 | false | false | 2024-09-03 | 2024-09-03 | 3 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Alpaca-Llama3.1.03-8B-Philos_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1.03-8B-Philos" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1.03-8B-Philos</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1.03-8B-Philos-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1.03-8B-Philos | 6e60f783f80f7d126b8e4f2b417e14dea63d2c4f | 20.29975 | apache-2.0 | 0 | 8 | true | false | false | false | 0.797523 | 0.388081 | 38.80814 | 0.495087 | 27.992549 | 0.129909 | 12.990937 | 0.278523 | 3.803132 | 0.42801 | 12.034635 | 0.335522 | 26.169105 | false | false | 2024-09-04 | 2024-09-04 | 3 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Alpaca-Llama3.1.04-8B-Philos_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1.04-8B-Philos" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1.04-8B-Philos</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1.04-8B-Philos-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1.04-8B-Philos | efd0c251373e1a2fa2bc8cead502c03ff6dc7c8b | 21.031577 | apache-2.0 | 0 | 8 | true | false | false | false | 0.765248 | 0.40844 | 40.843961 | 0.493001 | 27.963798 | 0.116314 | 11.63142 | 0.290268 | 5.369128 | 0.437219 | 13.685677 | 0.340259 | 26.695479 | false | false | 2024-09-05 | 2024-09-05 | 3 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Alpaca-Llama3.1.06-8B-Philos-dpo_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos-dpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos-dpo</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1.06-8B-Philos-dpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos-dpo | 3e76f190b505b515479cc25e92f8229c2b05159f | 21.829867 | apache-2.0 | 0 | 8 | true | false | false | false | 0.934774 | 0.486576 | 48.657562 | 0.488077 | 27.207177 | 0.128399 | 12.839879 | 0.297819 | 6.375839 | 0.393188 | 6.848437 | 0.361453 | 29.05031 | false | false | 2024-09-09 | 2024-09-09 | 5 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Alpaca-Llama3.1.07-8B-Philos-Math_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math | 0b2842bddfa6c308f67eb5a20daf04536a4e6d1a | 21.870165 | apache-2.0 | 0 | 8 | true | false | false | false | 0.90203 | 0.507908 | 50.790791 | 0.484702 | 26.901201 | 0.114048 | 11.404834 | 0.296141 | 6.152125 | 0.406302 | 7.854427 | 0.353059 | 28.117612 | false | false | 2024-09-10 | 2024-09-10 | 4 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Alpaca-Llama3.1.08-8B-C-R1-KTO-Reflection_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-C-R1-KTO-Reflection" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-C-R1-KTO-Reflection</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1.08-8B-C-R1-KTO-Reflection-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-C-R1-KTO-Reflection | dc900138b4406353b7e84251bc8649d70c16f13f | 20.882037 | apache-2.0 | 0 | 8 | true | false | false | false | 0.883974 | 0.395226 | 39.522578 | 0.495531 | 27.571611 | 0.123867 | 12.386707 | 0.299497 | 6.599553 | 0.404813 | 10.401563 | 0.359292 | 28.81021 | false | false | 2024-09-16 | 2024-09-16 | 6 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Alpaca-Llama3.1.08-8B-Philos-C-R1 | c57c786426123635baf6c8b4d30638d2053f4565 | 22.410483 | apache-2.0 | 0 | 8 | true | false | false | false | 0.909759 | 0.531638 | 53.163828 | 0.482793 | 26.763685 | 0.117825 | 11.782477 | 0.29698 | 6.263982 | 0.410302 | 8.454427 | 0.352311 | 28.034501 | false | false | 2024-09-13 | 2024-09-13 | 4 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-Llama-3.1-8B-Philos-Reflection_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Reflection" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Reflection</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Llama-3.1-8B-Philos-Reflection-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Reflection | 4b0b75d9235886e8a947c45b94f87c5a65a81467 | 20.389309 | apache-2.0 | 0 | 8 | true | false | false | false | 0.894943 | 0.359605 | 35.960474 | 0.489769 | 27.769796 | 0.129154 | 12.915408 | 0.307886 | 7.718121 | 0.395729 | 9.632813 | 0.355053 | 28.339243 | false | false | 2024-09-17 | 2024-09-17 | 5 | meta-llama/Meta-Llama-3.1-8B |
EpistemeAI2_Fireball-MathMistral-Nemo-Base-2407-v2dpo_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-MathMistral-Nemo-Base-2407-v2dpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-MathMistral-Nemo-Base-2407-v2dpo</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-MathMistral-Nemo-Base-2407-v2dpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-MathMistral-Nemo-Base-2407-v2dpo | 6b7d851c66359f39d16da6fbcf810b816dc6e4bc | 11.332218 | apache-2.0 | 1 | 11 | true | false | false | true | 1.881426 | 0.30972 | 30.972043 | 0.432764 | 21.145528 | 0.034743 | 3.47432 | 0.263423 | 1.789709 | 0.402958 | 8.969792 | 0.114777 | 1.641918 | false | false | 2024-08-21 | 2024-08-24 | 2 | unsloth/Mistral-Nemo-Base-2407-bnb-4bit |
EpistemeAI2_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-math_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-math" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-math</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-math-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-math | aa21037cf0984cb293facb69c41895e7fccb1340 | 22.677605 | apache-2.0 | 0 | 8 | true | false | false | false | 0.791683 | 0.551547 | 55.154656 | 0.480756 | 26.743767 | 0.132175 | 13.217523 | 0.30453 | 7.270694 | 0.36925 | 6.789583 | 0.342005 | 26.889406 | false | false | 2024-10-11 | 2024-10-12 | 3 | Removed |
EpistemeAI2_Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.005-128K-code-COT_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.005-128K-code-COT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.005-128K-code-COT</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.005-128K-code-COT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.005-128K-code-COT | cf8b99d4aa00c18fdaebfb24fa3c674ee6defa1a | 20.999994 | apache-2.0 | 0 | 8 | true | false | false | false | 0.800818 | 0.46332 | 46.331955 | 0.479083 | 26.400992 | 0.114804 | 11.480363 | 0.312081 | 8.277405 | 0.377438 | 5.013021 | 0.356466 | 28.496232 | false | false | 2024-10-11 | 2024-10-11 | 3 | Removed |
EpistemeAI2_Fireball-Phi-3-medium-4k-inst-Philos_bfloat16 | bfloat16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./EpistemeAI2/Fireball-Phi-3-medium-4k-inst-Philos" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">EpistemeAI2/Fireball-Phi-3-medium-4k-inst-Philos</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/EpistemeAI2__Fireball-Phi-3-medium-4k-inst-Philos-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | EpistemeAI2/Fireball-Phi-3-medium-4k-inst-Philos | 147715051102034fac98091e2a0cae6cade15ae0 | 29.172842 | apache-2.0 | 0 | 13 | true | false | false | true | 0.771814 | 0.531288 | 53.128809 | 0.617784 | 46.208873 | 0.140483 | 14.048338 | 0.332215 | 10.961969 | 0.413906 | 10.704948 | 0.459857 | 39.984116 | false | false | 2024-09-19 | 2024-09-20 | 1 | unsloth/phi-3-medium-4k-instruct-bnb-4bit |
Eric111_CatunaMayo_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./Eric111/CatunaMayo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Eric111/CatunaMayo</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Eric111__CatunaMayo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Eric111/CatunaMayo | 23337893381293975cbcc35f75b634954fbcefaf | 21.299155 | apache-2.0 | 0 | 7 | true | false | false | false | 0.550825 | 0.407416 | 40.741566 | 0.524364 | 33.299426 | 0.086103 | 8.610272 | 0.291946 | 5.592841 | 0.45399 | 15.348698 | 0.317819 | 24.202128 | true | false | 2024-02-15 | 2024-07-03 | 0 | Eric111/CatunaMayo |
Eric111_CatunaMayo-DPO_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./Eric111/CatunaMayo-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Eric111/CatunaMayo-DPO</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Eric111__CatunaMayo-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Eric111/CatunaMayo-DPO | 6bdbe06c10d57d152dd8a79a71edd8e30135b689 | 21.255121 | apache-2.0 | 0 | 7 | true | false | false | false | 0.554023 | 0.421454 | 42.145396 | 0.522399 | 33.089952 | 0.079305 | 7.930514 | 0.291946 | 5.592841 | 0.445031 | 14.66224 | 0.316988 | 24.109781 | true | false | 2024-02-21 | 2024-06-27 | 0 | Eric111/CatunaMayo-DPO |
Etherll_Chocolatine-3B-Instruct-DPO-Revised-Ties_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Chocolatine-3B-Instruct-DPO-Revised-Ties" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Chocolatine-3B-Instruct-DPO-Revised-Ties</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Chocolatine-3B-Instruct-DPO-Revised-Ties-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Chocolatine-3B-Instruct-DPO-Revised-Ties | 8a9c3d745e0805e769b544622b3f5c039abc9b07 | 24.402767 | 0 | 3 | false | false | false | false | 0.635497 | 0.372469 | 37.246949 | 0.541065 | 35.583343 | 0.128399 | 12.839879 | 0.323826 | 9.8434 | 0.464938 | 17.817187 | 0.397773 | 33.085845 | false | false | 2024-10-28 | 0 | Removed |
||
Etherll_Chocolatine-3B-Instruct-DPO-Revised-Ties-v2_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Phi3ForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Chocolatine-3B-Instruct-DPO-Revised-Ties-v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Chocolatine-3B-Instruct-DPO-Revised-Ties-v2</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Chocolatine-3B-Instruct-DPO-Revised-Ties-v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Chocolatine-3B-Instruct-DPO-Revised-Ties-v2 | 121b0831361743558e1a56fd89ae3d3c03272cc4 | 24.428163 | 0 | 3 | false | false | false | false | 0.631296 | 0.373993 | 37.399323 | 0.541065 | 35.583343 | 0.128399 | 12.839879 | 0.323826 | 9.8434 | 0.464938 | 17.817187 | 0.397773 | 33.085845 | false | false | 2024-10-29 | 0 | Removed |
||
Etherll_Herplete-LLM-Llama-3.1-8b_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Herplete-LLM-Llama-3.1-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Herplete-LLM-Llama-3.1-8b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Herplete-LLM-Llama-3.1-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Herplete-LLM-Llama-3.1-8b | b3829cf437216f099c031a9ab5e4c8ec974766dd | 19.588708 | 5 | 8 | false | false | false | true | 0.973685 | 0.467191 | 46.71915 | 0.501343 | 28.952591 | 0.027946 | 2.794562 | 0.286074 | 4.809843 | 0.386 | 6.683333 | 0.348155 | 27.572769 | false | false | 2024-08-24 | 2024-08-29 | 1 | Etherll/Herplete-LLM-Llama-3.1-8b (Merge) |
|
Etherll_Herplete-LLM-Llama-3.1-8b_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Herplete-LLM-Llama-3.1-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Herplete-LLM-Llama-3.1-8b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Herplete-LLM-Llama-3.1-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Herplete-LLM-Llama-3.1-8b | d1383d993fad005d515be4d815797019601c679f | 26.260139 | 5 | 8 | false | false | false | false | 0.854807 | 0.610598 | 61.059766 | 0.534725 | 33.206608 | 0.154834 | 15.483384 | 0.314597 | 8.612975 | 0.399052 | 8.614844 | 0.375249 | 30.583259 | false | false | 2024-08-24 | 2024-10-18 | 1 | Etherll/Herplete-LLM-Llama-3.1-8b (Merge) |
|
Etherll_Herplete-LLM-Llama-3.1-8b-Ties_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Herplete-LLM-Llama-3.1-8b-Ties" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Herplete-LLM-Llama-3.1-8b-Ties</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Herplete-LLM-Llama-3.1-8b-Ties-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Herplete-LLM-Llama-3.1-8b-Ties | 26.571056 | 0 | 8 | false | false | false | false | 0.862201 | 0.616368 | 61.63679 | 0.533798 | 33.07089 | 0.162387 | 16.238671 | 0.317114 | 8.948546 | 0.401719 | 8.948177 | 0.375249 | 30.583259 | false | false | 2024-10-03 | 2024-10-17 | 1 | Etherll/Herplete-LLM-Llama-3.1-8b-Ties (Merge) |
||
Etherll_Qwen2.5-7B-della-test_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Qwen2.5-7B-della-test" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Qwen2.5-7B-della-test</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Qwen2.5-7B-della-test-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Qwen2.5-7B-della-test | c2b2ffc38627e68e7b43a1b596dc16ee93c1c63b | 27.659468 | 1 | 7 | false | false | false | true | 1.385742 | 0.762497 | 76.249684 | 0.544733 | 35.546894 | 0 | 0 | 0.308725 | 7.829978 | 0.404698 | 8.98724 | 0.436087 | 37.343011 | false | false | 2024-11-01 | 2024-11-14 | 1 | Etherll/Qwen2.5-7B-della-test (Merge) |
|
Etherll_Qwen2.5-Coder-7B-Instruct-Ties_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Qwen2.5-Coder-7B-Instruct-Ties" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Qwen2.5-Coder-7B-Instruct-Ties</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Qwen2.5-Coder-7B-Instruct-Ties-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Qwen2.5-Coder-7B-Instruct-Ties | d8c1624a2fa60f05030e34a128af391b5d8be332 | 24.474445 | 0 | 7 | false | false | false | false | 1.197181 | 0.500539 | 50.053857 | 0.489514 | 28.008294 | 0.169184 | 16.918429 | 0.329698 | 10.626398 | 0.437281 | 13.426823 | 0.350316 | 27.812869 | false | false | 2024-09-30 | 2024-10-28 | 1 | Etherll/Qwen2.5-Coder-7B-Instruct-Ties (Merge) |
|
Etherll_Replete-LLM-V3-Llama-3.1-8b_float16 | float16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/Replete-LLM-V3-Llama-3.1-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/Replete-LLM-V3-Llama-3.1-8b</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__Replete-LLM-V3-Llama-3.1-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/Replete-LLM-V3-Llama-3.1-8b | e79849d72f70ef74677ed81a8885403973b2470c | 17.927882 | 5 | 8 | false | false | false | true | 0.789329 | 0.526292 | 52.629246 | 0.454338 | 22.902455 | 0.000755 | 0.075529 | 0.268456 | 2.46085 | 0.351646 | 2.055729 | 0.346991 | 27.443484 | false | false | 2024-08-24 | 2024-08-26 | 1 | Etherll/Replete-LLM-V3-Llama-3.1-8b (Merge) |
|
Etherll_SuperHermes_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Etherll/SuperHermes" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Etherll/SuperHermes</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Etherll__SuperHermes-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Etherll/SuperHermes | 7edd56cb37722d09b0334826e0532b223d334939 | 26.604602 | 1 | 8 | false | false | false | false | 0.750015 | 0.545902 | 54.590154 | 0.528953 | 32.840317 | 0.146526 | 14.652568 | 0.323826 | 9.8434 | 0.440042 | 14.938542 | 0.394864 | 32.762633 | false | false | 2024-10-27 | 2024-10-27 | 1 | Etherll/SuperHermes (Merge) |
|
Eurdem_Defne-llama3.1-8B_bfloat16 | bfloat16 | 🔶 fine-tuned on domain-specific datasets | 🔶 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Eurdem/Defne-llama3.1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Eurdem/Defne-llama3.1-8B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Eurdem__Defne-llama3.1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Eurdem/Defne-llama3.1-8B | 7832ba3066636bf4dab3e7d658c0b3ded12491ae | 25.095429 | llama3.1 | 3 | 8 | true | false | false | false | 1.7203 | 0.503612 | 50.361153 | 0.532098 | 32.822381 | 0.15861 | 15.861027 | 0.296141 | 6.152125 | 0.433094 | 13.536719 | 0.386553 | 31.83917 | false | false | 2024-07-29 | 2024-08-14 | 0 | Eurdem/Defne-llama3.1-8B |
FallenMerick_Chewy-Lemon-Cookie-11B_bfloat16 | bfloat16 | 🤝 base merges and moerges | 🤝 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./FallenMerick/Chewy-Lemon-Cookie-11B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FallenMerick/Chewy-Lemon-Cookie-11B</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/FallenMerick__Chewy-Lemon-Cookie-11B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | FallenMerick/Chewy-Lemon-Cookie-11B | 0f5d0d6d218b3ef034f58eba32d6fe7ac4c237ae | 22.018549 | cc-by-4.0 | 0 | 10 | true | false | false | false | 0.857274 | 0.487524 | 48.752421 | 0.525112 | 33.0143 | 0.05287 | 5.287009 | 0.279362 | 3.914989 | 0.454552 | 15.952344 | 0.326712 | 25.190233 | true | false | 2024-06-06 | 2024-06-27 | 1 | FallenMerick/Chewy-Lemon-Cookie-11B (Merge) |
Felladrin_Llama-160M-Chat-v1_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | LlamaForCausalLM | <a target="_blank" href="https://huggingface.co./Felladrin/Llama-160M-Chat-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Felladrin/Llama-160M-Chat-v1</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Felladrin__Llama-160M-Chat-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Felladrin/Llama-160M-Chat-v1 | e7f50665676821867ee7dfad32d0ca9fb68fc6bc | 4.101061 | apache-2.0 | 16 | 0 | true | false | false | true | 0.181581 | 0.157546 | 15.754642 | 0.303608 | 3.166756 | 0 | 0 | 0.25755 | 1.006711 | 0.366125 | 3.165625 | 0.113614 | 1.512633 | false | false | 2023-12-20 | 2024-07-23 | 1 | JackFram/llama-160m |
Felladrin_Minueza-32M-UltraChat_float16 | float16 | 💬 chat models (RLHF, DPO, IFT, ...) | 💬 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./Felladrin/Minueza-32M-UltraChat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Felladrin/Minueza-32M-UltraChat</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/Felladrin__Minueza-32M-UltraChat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | Felladrin/Minueza-32M-UltraChat | 28506b99c5902d2215eb378ec91d4226a7396c49 | 3.848727 | apache-2.0 | 4 | 0 | true | false | false | true | 0.168067 | 0.137563 | 13.756278 | 0.294148 | 2.43729 | 0 | 0 | 0.255872 | 0.782998 | 0.374187 | 4.640104 | 0.113281 | 1.475694 | false | false | 2024-02-27 | 2024-07-23 | 1 | Felladrin/Minueza-32M-Base |
FlofloB_100k_fineweb_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit_float16 | float16 | 🟩 continuously pretrained | 🟩 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co./FlofloB/100k_fineweb_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FlofloB/100k_fineweb_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/FlofloB__100k_fineweb_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | FlofloB/100k_fineweb_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit | ea6ceae8a6894f1c6ea3fe978846b2a66c3e369c | 7.871072 | apache-2.0 | 1 | 0 | true | false | false | true | 0.483694 | 0.308322 | 30.832192 | 0.332339 | 7.347825 | 0 | 0 | 0.269295 | 2.572707 | 0.330219 | 0.94401 | 0.149767 | 5.529699 | false | false | 2024-11-28 | 2024-11-29 | 3 | Qwen/Qwen2.5-0.5B |
FlofloB_10k_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit_float16 | float16 | 🟩 continuously pretrained | 🟩 | Original | MistralForCausalLM | <a target="_blank" href="https://huggingface.co./FlofloB/10k_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FlofloB/10k_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/FlofloB__10k_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | FlofloB/10k_continued_pretraining_Phi-3-mini-4k-instruct_Unsloth_merged_16bit | a2eb0460779e76bb511339bcc2545b4729c9d78e | 23.879918 | apache-2.0 | 1 | 16 | true | false | false | true | 0.487545 | 0.509731 | 50.973085 | 0.521499 | 32.6078 | 0.087613 | 8.761329 | 0.299497 | 6.599553 | 0.430958 | 13.569792 | 0.376912 | 30.767952 | false | false | 2024-11-22 | 2024-11-22 | 1 | unsloth/phi-3-mini-4k-instruct-bnb-4bit |
FlofloB_10k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit_float16 | float16 | 🟩 continuously pretrained | 🟩 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co./FlofloB/10k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FlofloB/10k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/FlofloB__10k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | FlofloB/10k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit | 2152657b389375f48fc5073413bba17835117bcc | 7.847811 | apache-2.0 | 1 | 0 | true | false | false | true | 0.508365 | 0.281544 | 28.154408 | 0.330552 | 7.530229 | 0 | 0 | 0.279362 | 3.914989 | 0.330219 | 1.477344 | 0.154089 | 6.0099 | false | false | 2024-11-25 | 2024-11-25 | 3 | Qwen/Qwen2.5-0.5B |
FlofloB_40k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit_float16 | float16 | 🟩 continuously pretrained | 🟩 | Original | Qwen2ForCausalLM | <a target="_blank" href="https://huggingface.co./FlofloB/40k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">FlofloB/40k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit</a> <a target="_blank" href="https://huggingface.co./datasets/open-llm-leaderboard/FlofloB__40k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a> | FlofloB/40k_continued_pretraining_Qwen2.5-0.5B-Instruct_Unsloth_merged_16bit | 64c61d9c777da56597a338afd7586cc4ad07d350 | 7.827703 | apache-2.0 | 1 | 0 | true | false | false | true | 0.481567 | 0.301578 | 30.157759 | 0.332461 | 7.53209 | 0 | 0 | 0.267617 | 2.348993 | 0.340823 | 1.536198 | 0.148521 | 5.391179 | false | false | 2024-11-25 | 2024-11-25 | 3 | Qwen/Qwen2.5-0.5B |