-
-
-
-
-
-
Inference Providers
Active filters:
llamacpp
DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters
kroonen/neural-chat-7b-v3-1-GGUF
7B
•
Updated
•
22
Druvith/mistralmed-7b-v1.5.gguf
7B
•
Updated
•
5
rxavier/Taurus-7B-1.0-GGUF
7B
•
Updated
•
54
BramVanroy/GEITje-7B-ultra-GGUF
7B
•
Updated
•
320
•
9
Vikhrmodels/it-5.3-fp16-32k-GGUF
8B
•
Updated
•
301
•
2
rubra-ai/Meta-Llama-3-8B-Instruct-GGUF
9B
•
Updated
•
106
•
4
Vikhrmodels/it-5.4-fp16-orpo-v2-GGUF
8B
•
Updated
•
284
•
4
Dracones/gemma-2-9b-it-GGUF
Text Generation
•
9B
•
Updated
•
8
Dracones/gemma-2-27b-it-GGUF
Text Generation
•
27B
•
Updated
•
64
Vikhrmodels/Vikhr-Gemma-2B-instruct-GGUF
Text Generation
•
3B
•
Updated
•
1.54k
•
16
flowaicom/Flow-Judge-v0.1-GGUF
Text Generation
•
4B
•
Updated
•
60
•
9
Vikhrmodels/Vikhr-Llama-3.2-1B-instruct-GGUF
Text Generation
•
1B
•
Updated
•
901
•
12
Vikhrmodels/Vikhr-Qwen-2.5-0.5B-instruct-GGUF
Text Generation
•
0.5B
•
Updated
•
245
•
6
Vikhrmodels/QVikhr-2.5-1.5B-Instruct-SMPO_GGUF
2B
•
Updated
•
235
•
9
Vikhrmodels/QVikhr-2.5-1.5B-Instruct-r_GGUF
2B
•
Updated
•
285
•
4
vicharai/ViCoder-html-32B-preview-GGUF
Text Generation
•
33B
•
Updated
•
68
•
4