Model converted specifically for NPU on Intel devices.

pip install --pre openvino==2025.2.0rc2 openvino-tokenizers==2025.2.0.0rc2 openvino-genai==2025.2.0.0rc2 --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly

NPU driver Version required : 32.0.100.4023

import openvino_genai as ov_genai

pipe = ov_genai.LLMPipeline("qwen3-0.6b-int4-ov-npu", device="NPU")
pipe.generate("The Sun is yellow because", max_new_tokens=500)
Downloads last month
10
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for bweng/qwen3-4b-int4-ov-npu

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Finetuned
(149)
this model

Collection including bweng/qwen3-4b-int4-ov-npu