File size: 1,047 Bytes
2c0b7f8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
---
license: apache-2.0
base_model:
- Qwen/Qwen3-4B
tags:
- autoround
---
This is [Qwen/Qwen3-4B](https://huggingface.co/Qwen/Qwen3-4B) quantized with [AutoRound](https://github.com/intel/auto-round/tree/main/auto_round) in 4-bit (symmetric + gptq format). The model has been created, tested, and evaluated by The Kaitchup.
The model is compatible with vLLM and Transformers.

More details in this article:
[How Well Does Qwen3 Handle 4-bit and 2-bit Quantization?](https://kaitchup.substack.com/p/how-well-does-qwen3-handle-4-bit)


![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b93e6bd6c468ac7536607e/3J5BLZXRl6eT8g11r1JDQ.png)


![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b93e6bd6c468ac7536607e/0wvK6MwnngzKA8m2qs7qS.png)

- **Developed by:** [The Kaitchup](https://kaitchup.substack.com/)
- **License:** Apache 2.0 license

## How to Support My Work
Subscribe to [The Kaitchup](https://kaitchup.substack.com/subscribe). This helps me a lot to continue quantizing and evaluating models for free.