view article Article Training and Finetuning Sparse Embedding Models with Sentence Transformers v5 By tomaarsen and 1 other • Jul 1 • 121
view reply Thanks for sharing!!!I might have spotted a minor mistake, should "To ensure full coverage of all domains in the non-thinking dataset" really be "To ensure full coverage of all domains in the thinking dataset"?
Running 3.23k 3.23k The Ultra-Scale Playbook 🌌 The ultimate guide to training LLM on large GPU Clusters