Anticipated Availability of GPTQModel Format Models (W4A16/W8A16)
#22 opened about 2 hours ago
by
X-SZM
Unable download the generated code
1
#21 opened 5 days ago
by
ssfarzad
Fixed π¨ GGUF Tool calling β MCP working β
1
#19 opened 15 days ago
by
xbruce22

Will there still be 32B dense models?
β
π
7
1
#18 opened 20 days ago
by
lingyezhixing
Upload Marginal adapttaion.pdf
#17 opened 22 days ago
by
thenunabdo
Please create 8-bit MLX - No-one has it anywhere...
#16 opened 23 days ago
by
Darkslayerofdark
Questions on FP8 inference, parallel requests, and context length with 4x H200s
2
#15 opened 25 days ago
by
sultan93
Does its api support formot?
#14 opened 25 days ago
by
Connde
Impressive Broad Knowledge
π
π
5
4
#12 opened 26 days ago
by
phil111
Thinking tokens issue
π
2
11
#9 opened 26 days ago
by
iyanello
Benchmarks for non-thinking mode
π
4
2
#8 opened 27 days ago
by
PSM24
Thankyou GLM Team for the wonderful MOE Model
π₯
6
#7 opened 27 days ago
by
Narutoouz

AWQ 4Bit / GPTQ with full precision gates and head? Please
8
#4 opened 28 days ago
by
chriswritescode

We Have Gemini At Home
4
#1 opened 28 days ago
by
MarinaraSpaghetti
