IntelligentEstate/Gambit-7B-Q4_K_M-GGUF

This model was converted to GGUF format from nvidia/AceReason-Nemotron-7B

Use with llama.cpp

Install llama.cpp through brew (works on Mac and Linux)

brew install llama.cpp

Invoke the llama.cpp server or the CLI.

Downloads last month
3
GGUF
Model size
7.62B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for IntelligentEstate/Gambit-7B-Q4_K_M-GGUF

Quantized
(15)
this model