Fizz 🏳️‍⚧️ PRO

Fizzarolli

AI & ML interests

None yet

Recent Activity

updated a dataset about 12 hours ago
allura-org/disco-elysium-conversations-raw
published a dataset about 13 hours ago
allura-org/disco-elysium-conversations-raw
liked a model about 17 hours ago
trashpanda-org/MS-24B-Mullein-v0
View all activity

Organizations

Zeus Labs's profile picture Social Post Explorers's profile picture dreamgen-preview's profile picture ShuttleAI's profile picture testing's profile picture Alfitaria's profile picture Allura's profile picture Estrogen's profile picture Smol Community's profile picture

Posts 2

view post
Post
2088
hi everyone!

i wanted to share an experiment i did with upcycling phi-3 mini into an moe recently.
while benchmarks are definitely within a margin of error and they performed similarly, i think it's an interesting base to try and see if you can improve phi's performance! (maybe looking into HuggingFaceFW/fineweb-edu could be interesting, i also left some other notes if anyone with more compute access wants to try it themselves)

check it out! Fizzarolli/phi3-4x4b-v1
view post
Post
2515
Is anyone looking into some sort of decentralized/federated dataset generation or classification by humans instead of synthetically?

From my experience with trying models, a *lot* of modern finetunes are trained on what amounts to, in essence, GPT-4 generated slop that makes everything sound like a rip-off GPT-4 (refer to i.e. the Dolphin finetunes). I have a feeling that this is a lot of the reason people haven't been quite as successful as Meta's instruct tunes of Llama 3.