Workflow
智能体系统如何「边做边学」?斯坦福团队探索在线优化的新范式
机器之心·2025-10-24 09:12

Core Insights - The article discusses the limitations of traditional methods for enabling intelligent agents to perform complex reasoning and tool usage, highlighting the need for a more scalable and adaptable approach [2][3][4] - The proposed AgentFlow framework integrates collaborative reasoning among multiple independent agent modules and introduces the Flow-GRPO algorithm for training, achieving significant performance improvements in various tasks [3][4][15] Group 1: Traditional Methods and Challenges - Traditional approaches to training language models for complex task reasoning either involve a single model handling both reasoning and tool usage or rely on static prompt-driven systems [11][14] - The first method struggles with stability and scalability in long-chain reasoning and dynamic environments, while the second lacks learning and adaptation capabilities [3][14] - The research team aimed to enable agent systems to learn and evolve through interaction, addressing the limitations of existing methods [14][15] Group 2: AgentFlow Framework - AgentFlow is a modular, tool-integrated intelligent agent system designed to overcome scalability and generalization limitations of current methods [15][27] - It features a planner that adapts in real-time during agent interactions, allowing for adaptive reasoning and robust tool-calling [15][19] - The framework demonstrates significant improvements in long-term planning, tool efficiency, and dynamic reasoning depth across various domains [4][15] Group 3: Flow-GRPO Algorithm - Flow-GRPO addresses the challenge of multi-turn credit assignment in reinforcement learning by broadcasting outcome rewards to each step, transforming complex multi-turn problems into manageable single-turn updates [19][20] - This method alleviates sparse reward issues and enhances training efficiency, providing a foundation for stable learning in complex reasoning tasks [20][27] Group 4: Experimental Results - AgentFlow was evaluated across ten benchmark tests, outperforming existing leading methods, including large proprietary models like GPT-4o [22][27] - Notable performance improvements include a 14.9% increase in knowledge retrieval, 14.0% in agentic reasoning, 14.5% in mathematical reasoning, and 4.1% in scientific reasoning [24][27] - The 7B parameter AgentFlow model surpassed the performance of 200B parameter models, demonstrating that effective system design can be more impactful than merely increasing model size [27][30] Group 5: Learning and Adaptation - The research indicates that online learning in real interaction environments is crucial for achieving efficient reasoning, as offline supervised training led to significant performance drops [27][30] - The system autonomously discovered new tool usage patterns, enhancing its ability to gather information through combined tool strategies [30][33] - AgentFlow's performance improves with increased reasoning steps without excessively extending average reasoning time, indicating effective task handling [33][35] Group 6: Conclusion and Future Potential - AgentFlow presents a novel approach to intelligent agent training, emphasizing continuous learning and adaptation over a single comprehensive model [36][37] - The work highlights the potential and imaginative possibilities within the field of agentic AI, despite the distance from research exploration to practical application [37]