Mistral 7B v0.1 - GGUF

This is a quantized model for mistralai/Mistral-7B-v0.1. Two quantization methods were used:

  • Q5_K_M: 5-bit, preserves most of the model's performance
  • Q4_K_M: 4-bit, smaller footprints and saves more memory

Description

This repo contains GGUF format model files for Mistral AI_'s Mistral 7B v0.1.

This model was quantized in Google Colab.

Downloads last month
20
GGUF
Model size
7.24B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for wenqiglantz/Mistral-7B-v0.1-GGUF

Quantized
(183)
this model