
howard-hou/RWKV-X
Updated
RWKV-X is a family of long-context models based on RWKV-7, enhanced with Sparse Attention and capable of handling context windows up to 64K tokens.
Note Model repository for Base and Instruct models (0.2B–3B parameters)