Michael Goin
mgoin
AI & ML interests
LLM inference optimization, compression, quantization, pruning, distillation
Recent Activity
updated
a model
about 2 hours ago
nm-testing/Llama-3_1-Nemotron-Ultra-253B-v1-FP8-dynamic
Organizations
mgoin's activity
how to deploy this model without internet connection
1
#1 opened 5 days ago
by
superahn
Why not FP8 with static and per-tensor quantization?
1
#2 opened 11 days ago
by
wanzhenchn
Address discrepancies in the languages supported by the Mistral Small 3.1 2503
1
2
#54 opened 16 days ago
by
fpaupier

Please update the chat template
1
#1 opened 16 days ago
by
stelterlab

FP8 Dynamic/W8A16 Quants Please
4
#44 opened 25 days ago
by
rjmehta
Problem hosting the model using vllm
3
4
#45 opened 25 days ago
by
ShaoServient
Remove image_processor_type
#1 opened about 2 months ago
by
pooya-davoodi-parasail
Remove image_processor_type
1
#1 opened about 2 months ago
by
pooya-davoodi-parasail
Remove image_processor_type
#2 opened about 2 months ago
by
pooya-davoodi-parasail
Use Qwen2VLImageProcessor for image_processor_type
5
#2 opened about 2 months ago
by
pooya-davoodi-parasail
Use Qwen2VLImageProcessor for image_processor_type
#3 opened about 2 months ago
by
pooya-davoodi-parasail
when i use vllm v0.7.2 to deploy r1 awq, i got empty content
13
#10 opened 2 months ago
by
bupalinyu
MLA is not supported with moe_wna16 quantization. Disabling MLA.
5
#7 opened 2 months ago
by
AMOSE
compressed-tensors MLA support requires fp8 activations and weights in group 'group_0',
2
#1 opened 3 months ago
by
samos123
How to load this model?
2
#1 opened 10 months ago
by
Frz614
Model does not run with VLLM
2
#3 opened 4 months ago
by
aswad546
Nice model, any info on scripts used to quantize?
1
#1 opened 4 months ago
by
RonanMcGovern
