llama3.2-3b-turkish-trained GGUF Quantized Models

Technical Details

  • Quantization Tool: llama.cpp
  • Version: version: 5155 (64082100)

Model Information

Available Files

πŸš€ Download πŸ”’ Type πŸ“ Description
Download Q5 K M 5-bit best (recommended HQ option)
Download Q8 0 8-bit maximum (overkill for most)
Download F16 Full precision (maximum accuracy)

πŸ’‘ Q4 K M provides the best balance for most use cases

Downloads last month
16
GGUF
Model size
3.21B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

5-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support