Intel NPU
Collection
Latest SOTA models supported on Intel NPU
•
6 items
•
Updated
Run Llama-3.2-1B optimized for Intel NPUs with nexaSDK.
Install nexaSDK and create a free account at sdk.nexa.ai
Activate your device with your access token:
nexa config set license '<access_token>'
Run the model on NPU in one line:
nexa infer NexaAI/llama3.2-1B-intel-npu
Llama-3.2-1B is the smallest model in the Llama 3.2 family, optimized for efficiency and ultra-lightweight deployment.
With just 1B parameters, it enables fast inference on resource-constrained environments while retaining strong instruction-following and multilingual capabilities for its size.
Input: Text prompts such as questions, instructions, or code snippets.
Output: Concise natural language responses, answers, or explanations.