Neo Dim
NeoDim
AI & ML interests
None yet
Recent Activity
liked
a model
2 days ago
bartowski/Confucius-o1-14B-GGUF
liked
a model
4 days ago
FuseAI/FuseO1-DeepSeekR1-QwQ-SkyT1-Flash-32B-Preview
liked
a model
4 days ago
bartowski/FuseO1-DeepSeekR1-QwQ-SkyT1-Flash-32B-Preview-GGUF
Organizations
None yet
NeoDim's activity
What is the prompt format?
13
#1 opened 11 months ago
by
siddhesh22
how did you convert `transformers.PreTrainedTokenizer` to ggml format?
1
#2 opened over 1 year ago
by
keunwoochoi
demo space
2
#4 opened over 1 year ago
by
matthoffner
Looks like the starchat-alpha-ggml-q4_1.bin is broken
8
#3 opened over 1 year ago
by
xhyi
missing tok_embeddings.weight error when trying to run with llama.cpp
2
#1 opened over 1 year ago
by
ultra2mh
Cannot run on llama.cpp and koboldcpp
3
#1 opened over 1 year ago
by
FenixInDarkSolo
Which inference repo is this quantized for?
3
#2 opened over 1 year ago
by
xhyi
Can the quantized model be loaded in gpu to have faster inference ?
6
#1 opened over 1 year ago
by
MohamedRashad
Can the quantized model be loaded in gpu to have faster inference ?
6
#1 opened over 1 year ago
by
MohamedRashad
Cannot run on llama.cpp and koboldcpp
3
#1 opened over 1 year ago
by
FenixInDarkSolo
Can the quantized model be loaded in gpu to have faster inference ?
6
#1 opened over 1 year ago
by
MohamedRashad