Safetensors
qwen2

Model Description

The RL-MemAgent-7B is a part of the MemAgent framework, which enables Large Language Models (LLMs) to process arbitrarily long texts through end-to-end Reinforcement Learning without altering their core architecture.

Usage

This model is ideal for tasks requiring the understanding and processing of very long documents, such as comprehensive question answering, summarizing extensive reports, or analyzing large codebases.

For detailed instructions on how to use, evaluate, and train models within the MemAgent framework, please refer to the main MemAgent GitHub repository.

Links

Citation

If you find this work useful, please consider citing our paper:

@article{yu2025memagent,
  title={MemAgent: Reshaping Long-Context LLM with Multi-Conv RL-based Memory Agent},
  author={Yu, Hongli and Chen, Tinghong and Feng, Jiangtao and Chen, Jiangjie and Dai, Weinan and Yu, Qiying and Zhang, Ya-Qin and Ma, Wei-Ying and Liu, Jingjing and Wang, Mingxuan and others},
  journal={arXiv preprint arXiv:2507.02259},
  year={2025}
}
Downloads last month
868
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for BytedTsinghua-SIA/RL-MemoryAgent-7B

Base model

Qwen/Qwen2.5-7B
Finetuned
(2446)
this model
Quantizations
2 models

Collection including BytedTsinghua-SIA/RL-MemoryAgent-7B