CardVault+ SmolVLM-500M GGUF Models
Available Models
Model File | Size | Quantization | Use Case |
---|---|---|---|
cardvault-500m-f16.gguf |
783MB | F16 (Base) | Maximum quality |
cardvault-500m-mmproj-f16.gguf |
191MB | F16 (Vision) | REQUIRED |
cardvault-500m-q8_0.gguf |
417MB | Q8_0 | Near-perfect quality |
cardvault-500m-q6_k.gguf |
399MB | Q6_K | Balanced |
cardvault-500m-q5_k_m.gguf |
311MB | Q5_K_M | Recommended |
cardvault-500m-q4_k_m.gguf |
290MB | Q4_K_M | Maximum compression |
Usage
# Download llama.cpp
git clone https://github.com/ggerganov/llama.cpp
cd llama.cpp && make
# Run inference (Q5_K_M recommended)
./main \
--model cardvault-500m-q5_k_m.gguf \
--mmproj cardvault-500m-mmproj-f16.gguf \
--image credit_card.jpg \
--prompt "Extract card information in JSON format"
โ ๏ธ Two-Component Architecture: Both text model + mmproj required!
- Downloads last month
- 23
Hardware compatibility
Log In
to view the estimation
4-bit
5-bit
6-bit
8-bit
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
Model tree for sugiv/cardvaultplus-500m-gguf
Base model
HuggingFaceTB/SmolLM2-360M
Quantized
HuggingFaceTB/SmolLM2-360M-Instruct
Quantized
HuggingFaceTB/SmolVLM-500M-Instruct
Quantized
sugiv/cardvaultplus-500m