---
language:
- en
- fr
- de
- es
- it
- pt
license: apache-2.0
library_name: transformers
inference: false
extra_gated_description: >-
If you want to learn more about how we process your personal data, please read
our Privacy Policy.
tags:
- vllm
base_model:
- Almawave/Velvet-14B
---
## DESCRIPTION
**This model does not represent the intended quality of the original product.**
To perform this quantization, we started with llama.cpp as our base, modifying the file convert_hf_to_gguf_update.py to support this model.
For modifying this file, we based our work on what was seen in the PR https://github.com/ggerganov/llama.cpp/pull/11716.
**Note:** As of today, llama.cpp does not support this model or this chat template https://github.com/ggerganov/llama.cpp/pull/11716.
## PROMPT FORMAT
Basic prompt format:
```
{prompt}
```
Prompt format with system message:
```
{system_prompt}
{prompt}
```
## DOWNLOAD
| Quant | Link |
| ------ | ---- |
| BF16 | [Velvet-14B-BF16](https://huggingface.co/SistInf/Velvet-14B-GGUF/blob/main/Velvet-14B-BF16.gguf) |
| F16 | [Velvet-14B-F16.gguf](https://huggingface.co/SistInf/Velvet-14B-GGUF/blob/main/Velvet-14B-F16.gguf) |
| Q4_K_M | [Velvet-14B-Q4_K_M](https://huggingface.co/SistInf/Velvet-14B-GGUF/blob/main/Velvet-14B-Q4_K_M.gguf) |
| Q4_K_S | [Velvet-14B-Q4_K_S](https://huggingface.co/SistInf/Velvet-14B-GGUF/blob/main/Velvet-14B-Q4_K_S.gguf) |
| Q5_K_M | [Velvet-14B-Q5_K_M](https://huggingface.co/SistInf/Velvet-14B-GGUF/blob/main/Velvet-14B-Q5_K_M.gguf) |
| Q6_K | [Velvet-14B-Q6_K.gguf](https://huggingface.co/SistInf/Velvet-14B-GGUF/blob/main/Velvet-14B-Q6_K.gguf) |
| Q8_0 | [Velvet-14B-Q8_0.gguf](https://huggingface.co/SistInf/Velvet-14B-GGUF/blob/main/Velvet-14B-Q8_0.gguf) |
Original Model: https://huggingface.co/Almawave/Velvet-14B
### License
Velvet-14B and Velvet-2B are made available under the Apache 2.0 license