Energy-Based Transformers are Scalable Learners and Thinkers Paper • 2507.02092 • Published 6 days ago • 37
Skywork-Reward-V2: Scaling Preference Data Curation via Human-AI Synergy Paper • 2507.01352 • Published 6 days ago • 46
KnutJaegersberg/Llama-3_3-Nemotron-Super-49B-GenRM-Multilingual-Q6_K-GGUF 50B • Updated 6 days ago • 37
KnutJaegersberg/Llama-3_3-Nemotron-Super-49B-GenRM-Multilingual-Q6_K-GGUF 50B • Updated 6 days ago • 37
nvidia/Llama-3_3-Nemotron-Super-49B-GenRM-Multilingual Text Generation • 50B • Updated 12 days ago • 67 • 6
Reward Models Collection Nemotron reward models. For use in RLHF pipelines and LLM-as-a-Judge • 8 items • Updated about 15 hours ago • 11