MiMo-VL-7B-SFT — 4-bit BitsAndBytes Quantized

This is a 4-bit quantized version of XiaomiMiMo/MiMo-VL-7B-SFT,
using the BitsAndBytes library.

Quantization reduces memory usage and makes it possible to run this model on consumer GPUs
(≤ 12 GB VRAM), at the cost of a small reduction in generation quality.


Quantization Details

  • Method: BitsAndBytes (bnb)
  • Precision: 4-bit (fp4)
  • Compute dtype: bfloat16
  • Double quantization: disabled
  • Format: safetensors
Downloads last month
228
Safetensors
Model size
8B params
Tensor type
F32
·
BF16
·
U8
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for NangWeiLun/MiMo-VL-7B-SFT-bnb-4bit-fp4

Unable to build the model tree, the base model loops to the model itself. Learn more.

Collection including NangWeiLun/MiMo-VL-7B-SFT-bnb-4bit-fp4