
TIGER-Lab/AceCoder-Qwen2.5-Coder-7B-Ins-V1.1
Updated
Note The state-of-the-art 7B reward model for code generation
Note The state-of-the-art 32B reward model for code generation
Note The first large-scale coding dataset with an average of 16 test cases per prompt, synthesized by GPT-4o-mini
Note DeepSeek-R1 style RL-tuned model with binary pass rate as the verifiable reward