UniGenBench-EvalModel-qwen-72b-v1

This model is tailored for offline T2I model evaluation on UniGenBench, which achieves an average accuracy of 94% compared to evaluations by Gemini 2.5 Pro.

Feel free to use this model to assess and compare the performance of your models.

image image

For further details, please refer to the following resources:

Citation

@article{UniGenBench,
  title={Pref-GRPO: Pairwise Preference Reward-based GRPO for Stable Text-to-Image Reinforcement Learning},
  author={Wang, Yibin and Li, Zhimin and Zang, Yuhang and Zhou, Yujie and Bu, Jiazi and Wang, Chunyu and Lu, Qinglin, and Jin, Cheng and Wang, Jiaqi},
  journal={arXiv preprint arXiv:2508.20751},
  year={2025}
}
Downloads last month
13
Safetensors
Model size
73.4B params
Tensor type
BF16
Β·
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for CodeGoat24/UniGenBench-EvalModel-qwen-72b-v1

Finetuned
(1)
this model

Dataset used to train CodeGoat24/UniGenBench-EvalModel-qwen-72b-v1

Collection including CodeGoat24/UniGenBench-EvalModel-qwen-72b-v1