Can LLM Reward Models Be Trusted? Master-RM Exposes and Fixes Their Weaknesses

Generative reward models, where large language models (LLMs) serve as evaluators, are gaining prominence in reinforcement learning with verifiable rewards (RLVR). These models are preferred over rule-based systems for tasks…

Continue Reading