Can Model Batch Infer By vLLM
#30 opened about 1 hour ago
by
BITDDD
Save VLLM model to local disk?
1
#29 opened 9 days ago
by
narai
OverflowError: out of range integral type conversion attempted
1
#28 opened 24 days ago
by
yangqingyou37
The different result with raw model and demo
2
#27 opened about 1 month ago
by
bluebluebluedd
Client Error : Can't load the model (missing config file)
1
#26 opened about 1 month ago
by
benhachem
Update README.md
2
#25 opened about 1 month ago
by
robertgshaw2
Update README.md
#24 opened about 1 month ago
by
narai
not support ollama
1
#23 opened about 2 months ago
by
nilzzz
cannot run model wit VLLM library - missting config.json file
2
#22 opened about 2 months ago
by
JBod
Add EXL2, INT8, and/or INT4 version of the model, PLEASE!
3
#21 opened about 2 months ago
by
Abdelhak
Cant run the Pixtral example inside readme because of library conflicts
2
#20 opened about 2 months ago
by
Valadaro
cuDNN error: CUDNN_STATUS_INTERNAL_ERROR
1
#19 opened 2 months ago
by
d3vnu77
Where is the gguf format?
1
#18 opened 2 months ago
by
RameshRajamani
how many languages supported?
2
#16 opened 2 months ago
by
xingwang1234
i am trying hf to gguf but there is no config
3
#15 opened 2 months ago
by
Batubatu
Updated README.md
1
#13 opened 2 months ago
by
drocks
Updated README.md
#12 opened 2 months ago
by
riaz
Use local image and quantise the model for low Gpu usage with solution
3
#11 opened 2 months ago
by
faizan4458
Fine-tuning
6
#10 opened 2 months ago
by
yukiarimo
Quantized Versions?
21
#9 opened 2 months ago
by
StopLockingDarkmode
Help
1
#8 opened 2 months ago
by
satvikahuja
Fix llm chat function call in README
#7 opened 2 months ago
by
ananddtyagi
Passing local images to chat (workaround).
1
#6 opened 2 months ago
by
averoo
MLX / MPS users out of luck and can't use this model with VLLM
1
#4 opened 2 months ago
by
kronosprime
Update README.md
#3 opened 2 months ago
by
pranay-ar