Qwen/Qwen2.5-VL-7B-Instruct (Quantized)

Description

This model is a quantized version of the original model Qwen/Qwen2.5-VL-7B-Instruct.

It's quantized using the BitsAndBytes library to 4-bit using the bnb-my-repo space.

Quantization Details

  • Quantization Type: int4
  • bnb_4bit_quant_type: nf4
  • bnb_4bit_use_double_quant: True
  • bnb_4bit_compute_dtype: bfloat16
  • bnb_4bit_quant_storage: uint8
Downloads last month
1
Safetensors
Model size
3.91B params
Tensor type
F32
BF16
U8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for medmekk/Qwen2.5-VL-7B-Instruct-2

Quantized
(26)
this model