Fizz 🏳️‍⚧️ PRO

Fizzarolli

AI & ML interests

None yet

Recent Activity

liked a dataset 2 days ago
allenai/tulu-3-sft-mixture
liked a model 2 days ago
fbellard/ts_server
liked a dataset 4 days ago
laion/LAION-DISCO-12M
View all activity

Organizations

Posts 2

view post
Post
1810
hi everyone!

i wanted to share an experiment i did with upcycling phi-3 mini into an moe recently.
while benchmarks are definitely within a margin of error and they performed similarly, i think it's an interesting base to try and see if you can improve phi's performance! (maybe looking into HuggingFaceFW/fineweb-edu could be interesting, i also left some other notes if anyone with more compute access wants to try it themselves)

check it out! Fizzarolli/phi3-4x4b-v1
view post
Post
2237
Is anyone looking into some sort of decentralized/federated dataset generation or classification by humans instead of synthetically?

From my experience with trying models, a *lot* of modern finetunes are trained on what amounts to, in essence, GPT-4 generated slop that makes everything sound like a rip-off GPT-4 (refer to i.e. the Dolphin finetunes). I have a feeling that this is a lot of the reason people haven't been quite as successful as Meta's instruct tunes of Llama 3.