Zheng Han
traphix
·
AI & ML interests
None yet
Recent Activity
Organizations
None yet
traphix's activity
Is there any accuracy results comparing to original DeepSeek-V3?
#6 opened 4 days ago
by
traphix
why "MLA is not supported with awq_marlin quantization. Disabling MLA." with 4090 * 32 (4 node / vllm 0.7.2)
3
#14 opened 4 days ago
by
FightLLM
Is there any accuracy results comparing to original DeepSeek-R1?
#15 opened 4 days ago
by
traphix
Has anyone evaluated the performance of the AWQ version of the model on benchmarks?
4
#8 opened 12 days ago
by
liuqianchao
skips the thinking process
11
#5 opened 16 days ago
by
muzizon
Deployment framework
27
#2 opened about 1 month ago
by
xro7
vllm support a100
17
#2 opened about 1 month ago
by
HuggingLianWang
Any plans to quantize Qwen/Qwen2.5-72B-Instruct to w8a8?
#1 opened 15 days ago
by
traphix
Can it run on A100/A800 with VLLM?
3
#1 opened 7 months ago
by
Parkerlambert123
Quantize DeepSeek-Coder-V2-Instruct to W8A8(INT8)?
#2 opened 6 months ago
by
traphix