This repository contains the model described in the paper RADLADS: Rapid Attention Distillation to Linear Attention Decoders at Scale.

Github repository: https://github.com/recursal/Monet

Downloads last month
104
Safetensors
Model size
8.09B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for recursal/QRWKV6-7B-Instruct

Quantizations
2 models

Collection including recursal/QRWKV6-7B-Instruct