add perplexity chart for v0.2 quants
Browse files- README.md +3 -1
- images/perplexity.png +3 -0
README.md
CHANGED
@@ -32,13 +32,15 @@ Also thanks to all the folks in the quanting and inferencing community on [Beave
|
|
32 |
## *UPDATED RECIPES*
|
33 |
Updated new better lower perplexity recipes and worlds smallest Kimi-K2-Instruct-smol-IQ1_KT at 219.375 GIB (1.835) BPW. Please ask any questions in [this discussion here](https://huggingface.co/ubergarm/Kimi-K2-Instruct-GGUF/discussions/4), thanks!
|
34 |
|
35 |
-
|
36 |
|
37 |
## Quant Collection
|
38 |
Compare with Perplexity of full size `Q8_0` 1016.623 GiB (8.504 BPW):
|
39 |
|
40 |
Final estimate: PPL = 2.9507 +/- 0.01468
|
41 |
|
|
|
|
|
42 |
### * v0.2 `IQ4_KS` 554.421 GiB (4.638 BPW)
|
43 |
Final estimate: PPL = 2.9584 +/- 0.01473
|
44 |
|
|
|
32 |
## *UPDATED RECIPES*
|
33 |
Updated new better lower perplexity recipes and worlds smallest Kimi-K2-Instruct-smol-IQ1_KT at 219.375 GIB (1.835) BPW. Please ask any questions in [this discussion here](https://huggingface.co/ubergarm/Kimi-K2-Instruct-GGUF/discussions/4), thanks!
|
34 |
|
35 |
+
Old versions still available as described in the dicsussion at tag/revision v0.1.
|
36 |
|
37 |
## Quant Collection
|
38 |
Compare with Perplexity of full size `Q8_0` 1016.623 GiB (8.504 BPW):
|
39 |
|
40 |
Final estimate: PPL = 2.9507 +/- 0.01468
|
41 |
|
42 |
+

|
43 |
+
|
44 |
### * v0.2 `IQ4_KS` 554.421 GiB (4.638 BPW)
|
45 |
Final estimate: PPL = 2.9584 +/- 0.01473
|
46 |
|
images/perplexity.png
ADDED
![]() |
Git LFS Details
|