Powerful ASR + diarization + speculative decoding with Hugging Face Inference Endpoints May 1, 2024 • 69
view article Article Train 400x faster Static Embedding Models with Sentence Transformers 3 days ago • 98
view article Article Introducing multi-backends (TRT-LLM, vLLM) support for Text Generation Inference 2 days ago • 41
Cosmos Tokenizer Collection A suite of image and video tokenizers • 13 items • Updated about 24 hours ago • 37
Llama 3.3 Collection This collection hosts the transformers and original repos of the Llama 3.3 • 1 item • Updated Dec 6, 2024 • 112
ESPnet-EZ: Python-only ESPnet for Easy Fine-tuning and Integration Paper • 2409.09506 • Published Sep 14, 2024 • 4
Lina-Speech: Gated Linear Attention is a Fast and Parameter-Efficient Learner for text-to-speech synthesis Paper • 2410.23320 • Published Oct 30, 2024 • 8
GTE models Collection General Text Embedding Models Released by Tongyi Lab of Alibaba Group • 19 items • Updated 28 days ago • 19
Smarter, Better, Faster, Longer: A Modern Bidirectional Encoder for Fast, Memory Efficient, and Long Context Finetuning and Inference Paper • 2412.13663 • Published about 1 month ago • 123
ModernBERT Collection Bringing BERT into modernity via both architecture changes and scaling • 3 items • Updated 30 days ago • 123
Prompting Depth Anything for 4K Resolution Accurate Metric Depth Estimation Paper • 2412.14015 • Published about 1 month ago • 12
Embedding Model Datasets Collection A curated subset of the datasets that work out of the box with Sentence Transformers: https://huggingface.co/datasets?other=sentence-transformers • 67 items • Updated Jul 3, 2024 • 95
FalconMamba 7B Collection This collection features the FalconMamba 7B base model, the instruction-tuned version, their 4-bit and GGUF variants, and the demo. • 15 items • Updated 10 days ago • 33
Bamba Collection Collection of Bamba - hybrid Mamba2 model architecture based models trained on open data • 8 items • Updated about 1 month ago • 18
Falcon3 Collection Falcon3 family of Open Foundation Models is a set of pretrained and instruct LLMs ranging from 1B to 10B parameters. • 40 items • Updated 10 days ago • 78
Gemma 2 2B Release Collection The 2.6B parameter version of Gemma 2. • 6 items • Updated Dec 13, 2024 • 78