Known stable releases of the miscii-1020 based models
Sthenno
sthenno
AI & ML interests
To contact me: [email protected]
Recent Activity
reacted
to
sometimesanotion's
post
with 👍
1 day ago
I've felt confident that 14B Qwen finetunes and merges could break the 42.0 average, and Arcee appears to have done exactly that with https://huggingface.co./arcee-ai/Virtuoso-Small-2. Congratulations to @arcee-ai!
Just two months ago, it was easy to think that 14B had plateaued, that you could have high IFEVAL or high MUSR/MATH/GPQA at 14B, but not both. That barrier is completely shattered. I see a pathway to even better, and Virtuoso Small 2 is a big part of why. Very impressive work. This community would expect no less from Arcee.
Just look at this graph! Keep in mind, my merges here build on the first Virtuoso Small, and *-DS merges build on DeepSeek R1. There are some impressive merges in the pipe!
replied to
sometimesanotion's
post
1 day ago
I've felt confident that 14B Qwen finetunes and merges could break the 42.0 average, and Arcee appears to have done exactly that with https://huggingface.co./arcee-ai/Virtuoso-Small-2. Congratulations to @arcee-ai!
Just two months ago, it was easy to think that 14B had plateaued, that you could have high IFEVAL or high MUSR/MATH/GPQA at 14B, but not both. That barrier is completely shattered. I see a pathway to even better, and Virtuoso Small 2 is a big part of why. Very impressive work. This community would expect no less from Arcee.
Just look at this graph! Keep in mind, my merges here build on the first Virtuoso Small, and *-DS merges build on DeepSeek R1. There are some impressive merges in the pipe!
reacted
to
sometimesanotion's
post
with 🚀
1 day ago
I've felt confident that 14B Qwen finetunes and merges could break the 42.0 average, and Arcee appears to have done exactly that with https://huggingface.co./arcee-ai/Virtuoso-Small-2. Congratulations to @arcee-ai!
Just two months ago, it was easy to think that 14B had plateaued, that you could have high IFEVAL or high MUSR/MATH/GPQA at 14B, but not both. That barrier is completely shattered. I see a pathway to even better, and Virtuoso Small 2 is a big part of why. Very impressive work. This community would expect no less from Arcee.
Just look at this graph! Keep in mind, my merges here build on the first Virtuoso Small, and *-DS merges build on DeepSeek R1. There are some impressive merges in the pipe!
Organizations
Collections
1
models
32
sthenno/tempesthenno-icy-0130
Text Generation
•
Updated
•
22
•
7
sthenno/tempesthenno-nuslerp-0124
Text Generation
•
Updated
•
88
•
4
sthenno/tempesthenno-0120
Updated
•
36
•
2
sthenno/tempesthenno-nuslerp-001
Text Generation
•
Updated
•
45
•
4
sthenno/tempesthenno-ppo-ckpt40
Updated
•
164
•
4
sthenno/tempesthenno-14b-nuslerp-0111
Updated
•
15
•
1
sthenno/tempesthenno-14b-0111
Updated
•
7
•
1
sthenno/tempesthenno-0126-ckpt150
Updated
•
23
•
1
sthenno/tempesthenno-hs2-rm
Updated
•
14
•
1
sthenno/miscii-1225-19b-preset
Text Generation
•
Updated
•
10
•
1
datasets
None public yet