🧠 Time-R1 Reinforced Model Weights

These are the official reinforcement learning (RL) fine-tuned model checkpoints for the paper: "Time Series Forecasting as Reasoning: A Slow-Thinking Approach with Reinforced LLMs".


πŸ“¦ Model Details

  • Base Model: Qwen2.5-7B
  • Tuning Framework: Verl + LLaMA Factory
  • Final Stage: Trained using GRIP (Group-based Relative Importance Policy optimization)
  • Objective: Multi-horizon time series forecasting with structured reasoning

πŸ“¦ Files Included

This model follows the standard Hugging Face transformers format and uses the efficient safetensors backend.

Time-R1/
β”œβ”€β”€ config.json
β”œβ”€β”€ generation_config.json
β”œβ”€β”€ model.safetensors.index.json
β”œβ”€β”€ model-00001-of-00004.safetensors
β”œβ”€β”€ model-00002-of-00004.safetensors
β”œβ”€β”€ model-00003-of-00004.safetensors
β”œβ”€β”€ model-00004-of-00004.safetensors
β”œβ”€β”€ tokenizer_config.json
β”œβ”€β”€ tokenizer.json
└── vocab.json

βœ… Fully compatible with Hugging Face transformers and AutoModelForCausalLM.

Downloads last month
1,459
Safetensors
Model size
7.62B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for ustc-zyt/Time-R1

Base model

Qwen/Qwen2.5-7B
Finetuned
(591)
this model
Quantizations
2 models

Dataset used to train ustc-zyt/Time-R1