RWKV7
Collection
RWKV7 models
•
3 items
•
Updated
•
1
This is RWKV-7 model under flash-linear attention format.
Install flash-linear-attention
and the latest version of transformers
before using this model:
pip install git+https://github.com/fla-org/flash-linear-attention
pip install 'transformers>=4.48.0'
You can use this model just as any other HuggingFace models:
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained('fla-hub/rwkv7-1.5B-world', trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained('fla-hub/rwkv7-1.5B-world', trust_remote_code=True)
This model is trained on the World v2.8 with a total of 1.0 trillion tokens.
lambada_openai
:
before conversion: ppl 4.13 acc 69.4%
after conversion: ppl 4.26 acc 68.8%
Q: safetensors metadata is none.
A: upgrade transformers to >=4.48.0: pip install 'transformers>=4.48.0'
Base model
BlinkDL/rwkv-7-world