Text Generation
Transformers
GGUF
Safetensors
mistral
quantized
2-bit
3-bit
4-bit precision
5-bit
6-bit
8-bit precision
GGUF
gemma
conversational
arxiv:2312.11805
arxiv:2009.03300
arxiv:1905.07830
arxiv:1911.11641
arxiv:1904.09728
arxiv:1905.10044
arxiv:1907.10641
arxiv:1811.00937
arxiv:1809.02789
arxiv:1911.01547
arxiv:1705.03551
arxiv:2107.03374
arxiv:2108.07732
arxiv:2110.14168
arxiv:2304.06364
arxiv:2206.04615
arxiv:1804.06876
arxiv:2110.08193
arxiv:2009.11462
arxiv:2101.11718
arxiv:1804.09301
arxiv:2109.07958
arxiv:2203.09509
Inference Endpoints
has_space
text-generation-inference
GGUF reuqest
#4
by
HR1777
- opened
Dear Maziyar,
I have discovered several intriguing models with exceptional performance that do not have a GGUF version. If possible, creating quininized versions of these models would be greatly appreciated. Please note that these models require authorization to access their respective repositories.
https://huggingface.co/xDAN-AI/xDAN-L1Mix-DeepThinking-v2
https://huggingface.co/xDAN-AI/xDAN-L1-Chat-v0.1
https://huggingface.co/xDAN-AI/xDAN-L1-moe-4x7b
https://huggingface.co/xDAN-AI/xDAN-L2-Chat-Performance-e2
https://huggingface.co/xDAN-AI/xDAN-L2-Chat-RL-v2
https://huggingface.co/xDAN-AI/xDAN-L1-moe-8x7b
Thank you so much. I hope they approve your access faster.