通过隐式奖励进行强化学习
Process Reinforcement through Implicit Rewards
February 3, 2025
作者: Ganqu Cui, Lifan Yuan, Zefan Wang, Hanbin Wang, Wendi Li, Bingxiang He, Yuchen Fan, Tianyu Yu, Qixin Xu, Weize Chen, Jiarui Yuan, Huayu Chen, Kaiyan Zhang, Xingtai Lv, Shuo Wang, Yuan Yao, Xu Han, Hao Peng, Yu Cheng, Zhiyuan Liu, Maosong Sun, Bowen Zhou, Ning Ding
cs.AI
摘要
在大型语言模型(LLMs)的推理时间扩展中,密集过程奖励已被证明是比稀疏的结果级奖励更有效的选择,特别是在需要复杂多步推理的任务中。虽然密集奖励对于强化学习(RL)LLMs也是一种吸引人的选择,因为它们的细粒度奖励有潜力解决一些结果级奖励固有的问题,比如训练效率和信用分配,但这种潜力主要仍未实现。这主要归因于在线训练过程奖励模型(PRMs)的挑战,收集高质量的过程标签成本过高,使其特别容易受到奖励欺骗的影响。为了解决这些挑战,我们提出了PRIME(通过隐式奖励进行过程强化),它通过隐式过程奖励,仅使用策略展开和结果标签来实现在线PRM更新。PRIME与各种优势函数结合,并放弃了现有方法所需的专门奖励模型训练阶段,大大降低了开发开销。我们在竞赛数学和编码方面展示了PRIME的有效性。从Qwen2.5-Math-7B-Base开始,PRIME在几个关键推理基准上平均提高了15.1%,超过了SFT模型。值得注意的是,我们的最终模型Eurus-2-7B-PRIME在七个推理基准上超过了Qwen2.5-Math-7B-Instruct模型,且只使用了其10%的训练数据。
English
Dense process rewards have proven a more effective alternative to the sparse
outcome-level rewards in the inference-time scaling of large language models
(LLMs), particularly in tasks requiring complex multi-step reasoning. While
dense rewards also offer an appealing choice for the reinforcement learning
(RL) of LLMs since their fine-grained rewards have the potential to address
some inherent issues of outcome rewards, such as training efficiency and credit
assignment, this potential remains largely unrealized. This can be primarily
attributed to the challenges of training process reward models (PRMs) online,
where collecting high-quality process labels is prohibitively expensive, making
them particularly vulnerable to reward hacking. To address these challenges, we
propose PRIME (Process Reinforcement through IMplicit rEwards), which enables
online PRM updates using only policy rollouts and outcome labels through
implict process rewards. PRIME combines well with various advantage functions
and forgoes the dedicated reward model training phrase that existing approaches
require, substantially reducing the development overhead. We demonstrate
PRIME's effectiveness on competitional math and coding. Starting from
Qwen2.5-Math-7B-Base, PRIME achieves a 15.1% average improvement across several
key reasoning benchmarks over the SFT model. Notably, our resulting model,
Eurus-2-7B-PRIME, surpasses Qwen2.5-Math-7B-Instruct on seven reasoning
benchmarks with 10% of its training data.Summary
AI-Generated Summary