Higher perplexity with Q5_K_S compared to legacy Q4_0?

#6
by stduhpf - opened

I did some perplexity measurements over the wikitext-2-raw-v1 dataset with iq3_xs, q4_0 and q5_k_s models (all from this repo), and surprisingly the q4_0 quant was significantly ahead.

model type PPL
IQ3_XS 10.0755 +/- 0.08024
Q4_0 9.5589 +/- 0.07527
Q5_K_S 9.8540 +/- 0.08016

That's not normal right? The Legacy format shouldn't perform better than the larger, more advanced format.

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment