iMatrix gguf quants of a newer finetune of Mixtral-8x22B

EdgeQuants still underway, IQ4XS version recommended. Make sure to combine/merge the parts back together before using

cat tessIQ4XS.gguf.part* > tessIQ4XS.gguf 

Then use with llama.cpp version from April 12 or older. April 13 release had massive changes and messed up inferene for MoE models

Downloads last month
4
GGUF
Model size
141B params
Architecture
llama
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for nisten/Tess-Mixtral-8x22B-imatrix-gguf

Quantized
(2)
this model