Custom GGUF quants of Meta’s Llama-3.2-Instruct's finetunes, where the Output Tensors are quantized to Q8_0 or F32 and the Embeddings are kept @F32
Joseph
Joseph717171
AI & ML interests
None yet
Recent Activity
liked
a model
1 day ago
google/gemma-3n-E4B-it
liked
a model
3 days ago
Menlo/Jan-nano-128k
new activity
4 days ago
unsloth/Jan-nano-GGUF:Please add the imatrix that you guys used for Jan-Nano (it's missing)