QuantFactory/Biggie-SmoLlm-0.4B-GGUF

This is quantized version of nisten/Biggie-SmoLlm-0.4B created using llama.cpp

Original Model Card

###Coherent Frankenstein of smolLm-0.36b upped to 0.4b

This took about 5 hours of semi-automated continuous merging to figure out the recipe. Model is smarter, and UNTRAINED. Uploaded it for training. Yet it performs well as is even quantized to 8bit. 8bit gguf included for testing.

wget https://huggingface.co/nisten/Biggie-SmoLlm-0.4B/resolve/main/Biggie_SmolLM_400M_q8_0.gguf
./llama-cli -ngl 99 -co --temp 0 -p "How to build a city on Mars via calculating Aldrin-Cycler orbits?" -m Biggie_SmolLM_400M_q8_0.gguf -cnv -fa --keep -1

image/png

Downloads last month
69
GGUF
Model size
401M params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for QuantFactory/Biggie-SmoLlm-0.4B-GGUF

Quantized
(18)
this model