AMD Instinct MI210 + vllm fail to run this model, any solutions please? Is there any other deepseek-r1-671b models that can run succesfully on AMD Instinct MI210 + vllm? Thanks!
5
#33 opened 16 days ago
by
luciagan
More stable startup command, not easy oom.
1
#31 opened about 1 month ago
by
Piekey
The awq quantization model may encounter garbled characters when performing inference on long texts.
9
#24 opened about 2 months ago
by
wx111
Add instructions to run R1-AWQ on SGLang
2
#22 opened about 2 months ago
by
ganler

requests get stuck when sending long prompts (already solved, but still don't know why?)
1
1
#18 opened 2 months ago
by
uv0xab
Is there any accuracy results comparing to original DeepSeek-R1?
2
#15 opened 2 months ago
by
traphix
Any one can run this model with SGlang framework?
5
#13 opened 2 months ago
by
muziyongshixin
Regarding the issue of inconsistent calculation of tokens
#12 opened 2 months ago
by
liguoyu3564
Max-Batch-Size, max-num-sequence, and fp_cache fp8_e4m3
#11 opened 2 months ago
by
BenFogerty