Anyone successful running any of these gguf models in vllm? All output I get is exclamations !!!
#6 opened 3 months ago
by
clawso
only 40k context accepted
#5 opened 5 months ago
by
rolead
4bit Q4_K_M Scored 63.57 on MMLU Pro single shot
π
1
#4 opened 5 months ago
by
xbruce22
Request support for MLX architecture
#3 opened 5 months ago
by
gowah
Tool issue Q8
1
#2 opened 5 months ago
by
sb6666
Any reason to not use this model over the 256K context model?
π
4
3
#1 opened 5 months ago
by
mallorbc