GRPO-VI-Qwen2-3B-RAG-RK3588-1.2.0
This version of GRPO-VI-Qwen2-3B-RAG has been converted to run on the RK3588 NPU using w8a8 quantization.
This model has been optimized with the following LoRA:
Compatible with RKLLM version: 1.2.0
Useful links:
Pretty much anything by these folks: marty1885 and happyme531
Converted using https://github.com/c0zaut/ez-er-rkllm-toolkit
Original Model Card for base model, GRPO-VI-Qwen2-3B-RAG, below:
AITeamVN/GRPO-VI-Qwen2-3B-RAG
- Downloads last month
- 4
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support