MiMo-VL-7B Quantize
Collection
XiaomiMiMo/MiMo-VL-7B Quantized Model
•
5 items
•
Updated
This is a 4-bit quantized version of XiaomiMiMo/MiMo-VL-7B-SFT,
using the BitsAndBytes library.
Quantization reduces memory usage and makes it possible to run this model on consumer GPUs
(≤ 12 GB VRAM), at the cost of a small reduction in generation quality.
fp4
) safetensors
Unable to build the model tree, the base model loops to the model itself. Learn more.