mmE5: Improving Multimodal Multilingual Embeddings via High-quality Synthetic Data Paper • 2502.08468 • Published 5 days ago • 12
InfiniteHiP: Extending Language Model Context Up to 3 Million Tokens on a Single GPU Paper • 2502.08910 • Published 5 days ago • 125
Scaling Pre-training to One Hundred Billion Data for Vision Language Models Paper • 2502.07617 • Published 6 days ago • 24
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling Paper • 2502.06703 • Published 7 days ago • 125
Gold-medalist Performance in Solving Olympiad Geometry with AlphaGeometry2 Paper • 2502.03544 • Published 12 days ago • 40
SmolLM2: When Smol Goes Big -- Data-Centric Training of a Small Language Model Paper • 2502.02737 • Published 13 days ago • 175
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models Paper • 2502.01061 • Published 14 days ago • 176
Streaming DiLoCo with overlapping communication: Towards a Distributed Free Lunch Paper • 2501.18512 • Published 18 days ago • 26
Qwen2.5-VL Collection Vision-language model series based on Qwen2.5 • 3 items • Updated 21 days ago • 345