John Leimgruber III
ubergarm
AI & ML interests
Open LLMs and Astrophotography image processing.
Organizations
None yet
ubergarm's activity
Observation: 4-bit quantization can't answer the Strawberry prompt
11
#2 opened 28 days ago
by
ThePabli
63.17 MMLU-Pro Computer Science with `Q8_0`
#2 opened about 1 month ago
by
ubergarm
Benchmarks worse than Qwen2.5-7B-Instruct on MMLU-Pro Computer Science in limited testing.
#1 opened about 1 month ago
by
ubergarm
Promising looking results on 24GB VRAM folks!
9
#3 opened about 2 months ago
by
ubergarm
Awesome model
6
#5 opened 2 months ago
by
dillfrescott
vram usage of each?
3
#1 opened 2 months ago
by
jasonden
Works good generating python on my 64GB RAM w/ 3090TI 24GB VRAM dev box
3
#2 opened 4 months ago
by
ubergarm
Chat template
3
#3 opened 4 months ago
by
sydneyfong
Can you please provide the command to change the context size?
5
#1 opened 4 months ago
by
yehiaserag
The first GGUF that works with long context on llama.cpp!
3
#1 opened 4 months ago
by
ubergarm
And where is the GGUF file itself?
12
#1 opened 4 months ago
by
Anonimus12345678902
Got it working in llama.cpp! Thanks!
1
#1 opened 4 months ago
by
ubergarm
Error loading model in llama.cpp ?
8
#1 opened 4 months ago
by
ubergarm
Prompt Format
4
#6 opened 6 months ago
by
JamesConley
Quantized model coming?
8
#3 opened 7 months ago
by
dnhkng
Output is empty
2
#3 opened 6 months ago
by
bingw5
The f16 with 32k ctx fits nicely in 24GB VRAM
5
#3 opened 7 months ago
by
ubergarm
AttributeError: 'generator' object has no attribute 'image_embeddings'
1
#26 opened 9 months ago
by
MohamedRashad