Dan Ofer

GrimSqueaker

AI & ML interests

Bioinformatics, Neurobiology, AutoML, Feature engineering, Proteins, NLP

Recent Activity

Organizations

None yet

GrimSqueaker's activity

view reply

I'd just start with modernBert large though, easier and strong base. Less faffing about. Also big vocab <3

view reply

They do PCA (prior to the zipf weighting) and explicitly state that they found that it improved perf.

view reply

Did you try potion/m2v as a starting point? (nvm modernbert, and it's much larger vocab)?

view reply

This is really cool! I'm surprised you do better than model2vec - is the difference really just the use of a (better) contrastive loss pretraining formula?