Zheng Han
traphix
·
AI & ML interests
None yet
Recent Activity
new activity about 13 hours ago
RedHatAI/Qwen3.5-122B-A10B-FP8-dynamic:Creation details? new activity about 14 hours ago
nivvis/Qwen3.5-122B-A10B-heretic-v2-FP8:Which framework was used for FP8 quantization? LLM-compressor? new activity 1 day ago
huihui-ai/Huihui-Qwen3-Coder-Next-abliterated:GPTQ quantizationOrganizations
None yet
Creation details?
#2 opened about 13 hours ago
by
traphix
Which framework was used for FP8 quantization? LLM-compressor?
2
#1 opened 1 day ago
by
traphix
GPTQ quantization
2
#2 opened about 1 month ago
by
ArtemSultanov
Question about weight_observer?
2
#1 opened 11 days ago
by
traphix
INT4 w4a16 quantinization?
➕ 1
#1 opened 21 days ago
by
traphix
Quantization code for int4(w4a16) ?
#6 opened 22 days ago
by
traphix
W4A16 quant
👍 2
4
#1 opened about 1 month ago
by
timroethig
Tokenizer you are loading with an incorrect regex pattern
1
#2 opened 3 months ago
by
traphix
Failed to find a kernel that can implement the WNA16 linear layer
#1 opened 3 months ago
by
traphix
vllm error: Extra inputs are not permitted
#1 opened 3 months ago
by
traphix
Can A100 run Qwen3-235B-A22B-Instruct-2507-NVFP4?
#1 opened 4 months ago
by
traphix
Error on 4 x L40s
➕ 2
1
#4 opened 6 months ago
by
traphix
I got ValueError
👀 2
10
#3 opened 6 months ago
by
spow12
How to run this model via vllm?
11
#2 opened 6 months ago
by
traphix
FP8 please
👀➕ 16
8
#18 opened 6 months ago
by
aliquis-pe
vllm v0.10.2 error
❤️ 1
#2 opened 6 months ago
by
traphix
VLLM compatibility?
9
#1 opened 6 months ago
by
aidendle94
Instruct or Thinking?
#1 opened 6 months ago
by
traphix