Core Argument - The over-reliance on Reinforcement Learning (RL) in AI development may be leading the industry away from achieving Artificial General Intelligence (AGI), as current models lack the ability to learn autonomously from experience like humans do [3][4]. Group 1: Skills Preconditioning Paradox - Current AI models depend on "pre-baked" skills, such as using Excel or browsing the web, which highlights their lack of general learning capabilities, indicating that AGI is not imminent [5]. - The approach of embedding specific skills into models contradicts the essence of human-like learning, which does not require extensive pre-training for every task [4][17]. Group 2: Insights from Robotics - The challenges in robotics stem from algorithmic issues rather than hardware limitations; if AI had human-like learning capabilities, robots would already be widely adopted without the need for repetitive training [6][13]. Group 3: Economic Implications of AI - The argument that "technology diffusion takes time" is seen as a self-comforting excuse; if models truly possessed human-like intelligence, they would be rapidly adopted by businesses due to lower risks and no training requirements [7][19]. - The disparity between the value created by global knowledge workers, amounting to trillions of dollars, and the significantly lower revenue generated by AI models indicates that these models have not yet reached the threshold to replace human workers [8][22]. Group 4: The Importance of Continual Learning - The key bottleneck for achieving AGI lies in the ability for "Continual Learning," rather than merely stacking RL computational power; true AGI may take another 10 to 20 years to realize [9][25]. - The process of solving the continual learning problem is expected to be gradual, similar to the evolution of context learning capabilities, and may not yield immediate breakthroughs [29][30].
大模型“缩放定律”悖论:RL(强化学习)越强,AGI(通用智能)越远?
硬AI·2025-12-24 08:10