GRPO-VI-Qwen2-3B-RAG-RK3588-1.2.0

This version of GRPO-VI-Qwen2-3B-RAG has been converted to run on the RK3588 NPU using w8a8 quantization.

This model has been optimized with the following LoRA:

Compatible with RKLLM version: 1.2.0

Useful links:

Official RKLLM GitHub

RockhipNPU Reddit

EZRKNN-LLM

Pretty much anything by these folks: marty1885 and happyme531

Converted using https://github.com/c0zaut/ez-er-rkllm-toolkit

Original Model Card for base model, GRPO-VI-Qwen2-3B-RAG, below:

AITeamVN/GRPO-VI-Qwen2-3B-RAG

Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support