Workflow
RLVMR框架
icon
Search documents
破解「长程智能体」RL训练难题,腾讯提出RLVMR框架,让7B模型「思考」比肩GPT-4o
机器之心· 2025-08-14 01:26
Core Viewpoint - The article discusses the development of the RLVMR framework by Tencent's Hunyuan AI Digital Human team, which aims to enhance the reasoning capabilities of AI agents by rewarding the quality of their thought processes rather than just the outcomes, addressing inefficiencies in long-horizon tasks and improving generalization abilities [4][26]. Group 1: Challenges in Current AI Agents - Many AI agents succeed in tasks but rely on luck and inefficient trial-and-error methods, leading to a lack of effective reasoning capabilities [2]. - The low-efficiency exploration problem arises as agents often engage in meaningless actions, resulting in high training costs and low reasoning efficiency [2]. - The generalization fragility issue occurs because strategies learned through guessing lack a logical foundation, making them vulnerable in new tasks [3]. Group 2: RLVMR Framework Introduction - RLVMR introduces a meta-reasoning approach that rewards good thinking processes, enabling end-to-end reinforcement learning for reasoning in long-horizon tasks [4][6]. - The framework allows agents to label their cognitive states, enhancing self-awareness and tracking their thought processes [7]. - A lightweight verification rule evaluates the quality of the agent's thinking in real-time, providing immediate rewards for good reasoning and penalizing ineffective habits [8]. Group 3: Experimental Results - The RLVMR-trained 7B model achieved a success rate of 83.6% on the most challenging L2 generalization tasks in ALFWorld and ScienceWorld, outperforming all previous state-of-the-art models [11]. - The number of actions required to solve tasks in complex environments decreased by up to 28.1%, indicating more efficient problem-solving paths [13]. - The training process showed faster convergence and more stable strategies, significantly alleviating the issue of ineffective exploration [13]. Group 4: Insights from RLVMR - The introduction of a reflection mechanism allows agents to identify problems and adjust strategies rather than blindly retrying, leading to a significant reduction in repeated actions and an increase in task success rates [19]. - Rewarding good reasoning habits establishes a flexible problem-solving framework that enhances generalization capabilities in unseen tasks [20][21]. - The two-phase training process of cold-start SFT followed by reinforcement learning aligns with cognitive principles, suggesting that teaching agents how to think before allowing them to learn from mistakes is more efficient [22][24]. Group 5: Conclusion and Future Outlook - RLVMR represents a paradigm shift from outcome-oriented to process-oriented training, effectively addressing the challenges of low-efficiency exploration and generalization fragility in long-horizon tasks [26]. - The ultimate goal is to develop AI agents capable of independent thinking and rational decision-making, moving beyond mere shortcut-seeking behaviors [26][27].