缩放定律(Scaling Law)
Search documents
大模型“缩放定律”悖论:RL(强化学习)越强,AGI(通用智能)越远?
硬AI· 2025-12-24 08:10
Core Argument - The over-reliance on Reinforcement Learning (RL) in AI development may be leading the industry away from achieving Artificial General Intelligence (AGI), as current models lack the ability to learn autonomously from experience like humans do [3][4]. Group 1: Skills Preconditioning Paradox - Current AI models depend on "pre-baked" skills, such as using Excel or browsing the web, which highlights their lack of general learning capabilities, indicating that AGI is not imminent [5]. - The approach of embedding specific skills into models contradicts the essence of human-like learning, which does not require extensive pre-training for every task [4][17]. Group 2: Insights from Robotics - The challenges in robotics stem from algorithmic issues rather than hardware limitations; if AI had human-like learning capabilities, robots would already be widely adopted without the need for repetitive training [6][13]. Group 3: Economic Implications of AI - The argument that "technology diffusion takes time" is seen as a self-comforting excuse; if models truly possessed human-like intelligence, they would be rapidly adopted by businesses due to lower risks and no training requirements [7][19]. - The disparity between the value created by global knowledge workers, amounting to trillions of dollars, and the significantly lower revenue generated by AI models indicates that these models have not yet reached the threshold to replace human workers [8][22]. Group 4: The Importance of Continual Learning - The key bottleneck for achieving AGI lies in the ability for "Continual Learning," rather than merely stacking RL computational power; true AGI may take another 10 to 20 years to realize [9][25]. - The process of solving the continual learning problem is expected to be gradual, similar to the evolution of context learning capabilities, and may not yield immediate breakthroughs [29][30].
南财快评|如何看待美股AI估值争议?
2 1 Shi Ji Jing Ji Bao Dao· 2025-11-21 11:28
Core Viewpoint - Nvidia's third-quarter earnings report exceeded expectations, with revenue of $57.01 billion and net profit of $31.91 billion, reflecting year-on-year growth of 62% and 65% respectively, which may alleviate concerns about AI industry valuations in the stock market [2] Group 1: Financial Performance - Nvidia's Q3 revenue was $57.01 billion, surpassing market expectations of $54.92 billion, and showing a year-on-year increase of 62% [2] - The net profit for the same period was $31.91 billion, marking a significant year-on-year increase of 65% [2] Group 2: Market Dynamics - The current AI boom in the U.S. is largely driven by supply-side investments from major tech companies like Microsoft, Google, and Meta, which are heavily investing in Nvidia's GPUs to build computing power centers [2] - There are concerns that the capital expenditures for AI infrastructure are exceeding current actual demand, drawing parallels to the internet bubble of 2000 [3] Group 3: Technological Evolution - Historical tech revolutions often experience bubbles as a necessary phase, with capital flowing in before technology matures, which can lead to resource misallocation but also provides funding for technological advancements [3] - The accumulation of computing power globally may be a necessary step towards achieving Artificial General Intelligence (AGI) [3] Group 4: Future Challenges - The tech giants are entering a challenging phase where the expectations for technology commercialization must catch up with rising anticipations [4] - Investors are increasingly demanding tangible revenue and profit margins rather than just optimistic future projections, indicating a shift in focus from merely accumulating computing power to demonstrating real profitability [4] Group 5: Valuation Concerns - A potential resolution to the current valuation debate could involve a "time for space" process, where gradual technology application leads to more reasonable valuations, requiring patience from market investors [5]
我们扒完了 GPT-5 全网爆料,奥特曼和 OpenAI 这次的饼真不好画了
3 6 Ke· 2025-08-05 10:39
Core Insights - OpenAI's marketing strategy for GPT-5 has faced criticism for being overly hyped without substantial product information [3][5][7] - The anticipated release of GPT-5 has been delayed, with various rumors about its launch date circulating since last year [8][23][26] - GPT-5 is expected to feature significant upgrades in multi-modal capabilities, software engineering, and AI agent functionalities [9][10][12][15] Group 1: Marketing and Anticipation - OpenAI has been criticized for its marketing tactics, which include frequent but vague updates about GPT-5, leading to public skepticism [3][5][7] - The excitement around GPT-5 has diminished despite OpenAI's continuous benchmark improvements, indicating a shift in marketing strategy to maintain interest [7][8] - Speculation about the release date has intensified, with predictions pointing towards a launch in early August 2025 [23][26][31] Group 2: Technical Advancements - GPT-5's core upgrades include a unified foundational and reasoning model, enhancing its performance in practical applications [9][12] - The model is expected to achieve "complete multi-modal" capabilities, allowing it to process and generate various types of media more effectively [10][11] - Significant improvements in software engineering capabilities will enable GPT-5 to modify and maintain complex enterprise-level codebases, challenging competitors like Anthropic [12][14] Group 3: AI Agent and Reasoning - GPT-5 is designed to execute complex multi-step tasks with reduced human supervision, marking a step towards autonomous AI agents [15][18] - The introduction of a "universal verifier" technology aims to enhance the model's ability to evaluate responses in subjective domains, improving its performance in creative writing and strategy analysis [16][18] - The model's architecture may include a smart routing system that dynamically selects the most suitable model for user queries based on complexity [22] Group 4: Development Challenges - The development of GPT-5 has faced challenges, including the limitations of traditional pre-training methods and the need for a new approach to scaling [40][41] - Internal projects like "Orion" did not meet expectations, leading to a pivot in strategy towards enhancing reasoning capabilities [40][41] - The successful development of the Q* technology has significantly improved the model's reasoning abilities, allowing it to tackle previously unseen problems [41][42] Group 5: Competitive Landscape - OpenAI's upcoming release of GPT-5 is crucial for regaining its competitive edge in areas like programming, where it has lost ground to competitors [50] - The partnership with Microsoft is aimed at leveraging OpenAI's technology to enhance Microsoft's own products while maintaining flexibility in their collaboration [32][35] - The competitive landscape is intensifying, with other companies like Anthropic and Google DeepMind also preparing to launch their advanced models [48][50]