ubergarm commited on
Commit
5dc7c55
·
1 Parent(s): 694ec54

add perplexity chart for v0.2 quants

Browse files
Files changed (2) hide show
  1. README.md +3 -1
  2. images/perplexity.png +3 -0
README.md CHANGED
@@ -32,13 +32,15 @@ Also thanks to all the folks in the quanting and inferencing community on [Beave
32
  ## *UPDATED RECIPES*
33
  Updated new better lower perplexity recipes and worlds smallest Kimi-K2-Instruct-smol-IQ1_KT at 219.375 GIB (1.835) BPW. Please ask any questions in [this discussion here](https://huggingface.co/ubergarm/Kimi-K2-Instruct-GGUF/discussions/4), thanks!
34
 
35
- Look there for graph with new values. I'll update the model card after the dust has settled. Old versions still available as described in the dicsussion.
36
 
37
  ## Quant Collection
38
  Compare with Perplexity of full size `Q8_0` 1016.623 GiB (8.504 BPW):
39
 
40
  Final estimate: PPL = 2.9507 +/- 0.01468
41
 
 
 
42
  ### * v0.2 `IQ4_KS` 554.421 GiB (4.638 BPW)
43
  Final estimate: PPL = 2.9584 +/- 0.01473
44
 
 
32
  ## *UPDATED RECIPES*
33
  Updated new better lower perplexity recipes and worlds smallest Kimi-K2-Instruct-smol-IQ1_KT at 219.375 GIB (1.835) BPW. Please ask any questions in [this discussion here](https://huggingface.co/ubergarm/Kimi-K2-Instruct-GGUF/discussions/4), thanks!
34
 
35
+ Old versions still available as described in the dicsussion at tag/revision v0.1.
36
 
37
  ## Quant Collection
38
  Compare with Perplexity of full size `Q8_0` 1016.623 GiB (8.504 BPW):
39
 
40
  Final estimate: PPL = 2.9507 +/- 0.01468
41
 
42
+ ![Perplexity Chart](images/perplexity.png "Chart showing Perplexity improving as BPW increases.")
43
+
44
  ### * v0.2 `IQ4_KS` 554.421 GiB (4.638 BPW)
45
  Final estimate: PPL = 2.9584 +/- 0.01473
46
 
images/perplexity.png ADDED

Git LFS Details

  • SHA256: f2f7c2e92a9ebd8f52d77ba8d04ad2bfce3d5763398f4f7450e06ed262ac5374
  • Pointer size: 131 Bytes
  • Size of remote file: 204 kB