Star Attention: Efficient LLM Inference over Long Sequences Paper • 2411.17116 • Published 5 days ago • 41
Star Attention: Efficient LLM Inference over Long Sequences Paper • 2411.17116 • Published 5 days ago • 41
Star Attention: Efficient LLM Inference over Long Sequences Paper • 2411.17116 • Published 5 days ago • 41 • 2
nGPT: Normalized Transformer with Representation Learning on the Hypersphere Paper • 2410.01131 • Published Oct 1 • 9
HelpSteer2: Open-source dataset for training top-performing reward models Paper • 2406.08673 • Published Jun 12 • 16
NeMo-Aligner: Scalable Toolkit for Efficient Model Alignment Paper • 2405.01481 • Published May 2 • 25
Every child should have parents: a taxonomy refinement algorithm based on hyperbolic term embeddings Paper • 1906.02002 • Published Jun 5, 2019 • 1
Damage Control During Domain Adaptation for Transducer Based Automatic Speech Recognition Paper • 2210.03255 • Published Oct 6, 2022 • 1
RULER: What's the Real Context Size of Your Long-Context Language Models? Paper • 2404.06654 • Published Apr 9 • 34
Damage Control During Domain Adaptation for Transducer Based Automatic Speech Recognition Paper • 2210.03255 • Published Oct 6, 2022 • 1
Every child should have parents: a taxonomy refinement algorithm based on hyperbolic term embeddings Paper • 1906.02002 • Published Jun 5, 2019 • 1
RULER: What's the Real Context Size of Your Long-Context Language Models? Paper • 2404.06654 • Published Apr 9 • 34
Nemotron 3 8B Collection The Nemotron 3 8B Family of models is optimized for building production-ready generative AI applications for the enterprise. • 5 items • Updated Oct 1 • 46