Scaling Law
Search documents
“推理模型还处于RNN的阶段”——李建忠对话GPT-5与Transformer发明者Lukasz Kaiser实录
AI科技大本营· 2025-10-10 09:52
Core Insights - The dialogue emphasizes the evolution of AI, particularly the transition from language models to reasoning models, highlighting the need for a new level of innovation akin to the Transformer architecture [1][2][4]. Group 1: Language and Intelligence - Language plays a crucial role in AI development, with the emergence of large language models marking a significant leap in AI intelligence [6][8]. - The understanding of language as a time-dependent sequence is essential for expressing intelligence, as it allows for continuous generation and processing of information [7][9]. - Current models exhibit the ability to form abstract concepts, similar to human learning processes, despite criticisms of lacking true understanding [9][10]. Group 2: Multimodal and World Models - The pursuit of unified models for different modalities is ongoing, with current models like GPT-4 already demonstrating multimodal capabilities [12][13]. - There is skepticism regarding the sufficiency of language models alone for achieving AGI, with some experts advocating for world models that learn physical world rules through observation [14][15]. - Improvements in model architecture and data quality are necessary to bridge the gap between language and world models [15][16]. Group 3: AI Programming - AI programming is seen as a significant application of language models, with potential shifts towards natural language-based programming [17][19]. - Two main perspectives on the future of AI programming exist: one advocating for AI-native programming and the other for AI as a copilot, suggesting a hybrid approach [18][20]. Group 4: Agent Models and Generalization - The concept of agent models is discussed, with challenges in generalization to new tasks being a key concern [21][22]. - The effectiveness of agent systems relies on the ability to learn from interactions and utilize external tools, which is currently limited [22][23]. Group 5: Scaling Laws and Computational Limits - The scaling laws in AI development are debated, with concerns about over-reliance on computational power potentially overshadowing algorithmic advancements [24][25]. - The economic limits of scaling models are acknowledged, suggesting a need for new architectures beyond the current paradigms [25][28]. Group 6: Embodied Intelligence - The slow progress in embodied intelligence, particularly in robotics, is attributed to data scarcity and fundamental differences between bits and atoms [29][30]. - Future models capable of understanding and acting in the physical world are anticipated, requiring advancements in multimodal training [30][31]. Group 7: Reinforcement Learning - The shift towards reinforcement learning-driven reasoning models is highlighted, with potential for significant scientific discoveries [32][33]. - The current limitations of RL training methods are acknowledged, emphasizing the need for further exploration and improvement [34]. Group 8: AI Organization and Collaboration - The development of next-generation reasoning models is seen as essential for achieving large-scale agent collaboration [35][36]. - The need for more parallel processing and effective feedback mechanisms in agent systems is emphasized to enhance collaborative capabilities [36][37]. Group 9: Memory and Learning - The limitations of current models' memory capabilities are discussed, with a focus on the need for more sophisticated memory mechanisms [37][38]. - Continuous learning is identified as a critical area for future development, with ongoing efforts to integrate memory tools into models [39][40]. Group 10: Future Directions - The potential for next-generation reasoning models to achieve higher data efficiency and generate innovative insights is highlighted [41].
OpenAI奥特曼认错:我天生不适合管理公司
量子位· 2025-10-09 07:03
Core Insights - OpenAI is pursuing three main goals: to become a personal AI subscription service, to build large-scale infrastructure, and to achieve a truly useful AGI (Artificial General Intelligence) [2][4][29] - The recent launch of Sora 2 and various investment collaborations, including partnerships with AMD and Nvidia, indicate a strategic shift towards aggressive infrastructure investment [1][29] Group 1: OpenAI's Strategic Goals - OpenAI aims to become a personal AI subscription service, necessitating the construction of vast infrastructure to support this vision [4][29] - The ultimate mission is to create AGI that is genuinely beneficial to humanity, which requires a multifaceted approach beyond traditional business models [4][8] - OpenAI's infrastructure is currently intended for internal use, with future possibilities for external applications remaining uncertain [5][29] Group 2: Sora's Role in AGI Development - Despite skepticism about Sora's relevance to AGI, OpenAI's CEO believes that developing a "truly outstanding world model" through Sora will be crucial for AGI [10][11] - The resources allocated to Sora are relatively small compared to OpenAI's overall computational capacity, emphasizing a balanced approach to innovation and research [13][29] - Sora is seen as a way to engage society with upcoming technological advancements, particularly in video models, which resonate more emotionally than text [16][29] Group 3: Future Interactions and AI Capabilities - OpenAI envisions future interaction interfaces that go beyond basic chat, incorporating real-time video rendering and context-aware hardware [19][21] - The concept of the Turing Test is evolving, with the new benchmark being AI's ability to conduct scientific research, which OpenAI anticipates will happen within two years [21][22] - OpenAI's confidence in its research roadmap and the economic value it can generate has led to a commitment to aggressive infrastructure investments [29][31] Group 4: Leadership and Management Philosophy - OpenAI's CEO acknowledges a preference for an investor role over management, citing challenges in handling organizational dynamics and operational details [41][42] - The transition from an investor to a CEO role has been described as both challenging and rewarding, providing insights into groundbreaking work in AI [41][43] - The future of AI development is closely tied to energy availability, with a call for more efficient energy solutions to support AI advancements [44]
听说,大家都在梭后训练?最佳指南来了
机器之心· 2025-10-09 02:24
Core Insights - The article emphasizes the shift in focus from pre-training to post-training in large language models (LLMs), highlighting the diminishing returns of scaling laws as model sizes reach hundreds of billions of parameters [2][3][11]. Group 1: Importance of Post-Training - Post-training is recognized as a crucial phase for enhancing the reasoning capabilities of models like OpenAI's series, DeepSeek R1, and Google Gemini, marking it as a necessary step towards advanced intelligence [3][11]. - The article introduces various innovative post-training methods such as Reinforcement Learning from Human Feedback (RLHF), Reinforcement Learning from AI Feedback (RLAIF), and Reinforcement Learning with Verifiable Rewards (RLVR) [2][3][12]. Group 2: Transition from Pre-Training to Post-Training - The evolution from pre-training to instruction fine-tuning is discussed, where foundational models are trained on large datasets to predict the next token, but often lack practical utility in real-world applications [7][8]. - Post-training aims to align model behavior with user expectations, focusing on quality over quantity in the datasets used, which are typically smaller but more refined compared to pre-training datasets [11][24]. Group 3: Supervised Fine-Tuning (SFT) - Supervised Fine-Tuning (SFT) is described as a process that transforms a pre-trained model into one that can follow user instructions effectively, relying on high-quality instruction-answer pairs [21][24]. - The quality of the SFT dataset is critical, as even a small number of low-quality samples can negatively impact the model's performance [25][26]. Group 4: Reinforcement Learning Techniques - Reinforcement Learning (RL) is highlighted as a complex yet effective method for model fine-tuning, with various reward mechanisms such as RLHF, RLAIF, and RLVR being employed to enhance model performance [39][41]. - The article outlines the importance of reward models in RLHF, which are trained using human preference data to guide model outputs [44][46]. Group 5: Evaluation of Post-Training Models - The evaluation of post-training models is multifaceted, requiring a combination of automated and human assessments to capture various quality aspects [57][58]. - Automated evaluations are cost-effective and quick, while human evaluations provide a more subjective quality measure, especially for nuanced tasks [59][60].
“大就是好”,但技术男阿里云并不执著“上头条”
Guan Cha Zhe Wang· 2025-09-29 09:46
Core Viewpoint - Alibaba's CEO, Wu Yongming, delivered a notable presentation at the Yunqi Conference, which led to a significant 9.16% increase in Alibaba's stock price, indicating strong investor sentiment despite a generally cautious market environment [1][3]. Group 1: Company Developments - Wu Yongming highlighted that large models will dominate software as the next-generation operating system, and Alibaba Cloud plans to invest further in AI infrastructure beyond its existing 380 billion yuan commitment over three years [3]. - Alibaba Cloud's Qwen3-Max model has achieved significant advancements, including an increase in pre-training data from 18 terabytes to 36 terabytes, and a focus on scaling laws to enhance model performance [6][10]. - The company has positioned itself as a leader in the AI cloud market, with a reported 35.8% market share, significantly ahead of competitors [16][22]. Group 2: Competitive Landscape - The competition in the AI cloud sector is intensifying, particularly with ByteDance's Volcano Engine, which has captured a 49.2% market share in model-as-a-service (MaaS) [16][18]. - Despite the competitive pressure, Alibaba Cloud has maintained a strong position, with over 53% of Fortune 500 companies using its services for generative AI [16][22]. - The market dynamics are shifting, with a trend towards self-deployment of models on Alibaba Cloud rather than relying solely on API calls, which may not be fully reflected in market share statistics [16][22]. Group 3: Technological Innovations - Alibaba Cloud has made significant strides in AI infrastructure, including the development of a new AI chip that approaches NVIDIA's capabilities and a high-performance network architecture that supports large-scale GPU interconnectivity [25][27]. - The company is focusing on a comprehensive stack for AI infrastructure, which positions it well in the context of increasing domestic demand for AI capabilities [27]. - Innovations in model architecture, such as the introduction of the Qwen3-Next model with a sparse MoE architecture, demonstrate Alibaba's commitment to advancing AI technology [6][10].
人形与具身智能产业何以叩响“Scaling Law”之门?
机器人大讲堂· 2025-09-24 11:09
Core Viewpoint - The humanoid robot industry is at a critical transformation point, moving from early "theme speculation" to "pre-investment in industrial trends" as companies like Tesla and Figure begin small-scale production. The industry's non-linear growth hinges on breakthroughs in hardware cost reduction and advancements in intelligent robotics [1][3]. Group 1: Current Industry Landscape - The core contradiction in humanoid robotics is not about "whether to ship" but rather "whether to form a sustainable industrial flywheel." By the end of 2024 and early 2025, many domestic companies have completed deliveries of hundreds to thousands of units, primarily in research, education, and display sectors [1][3]. - Initial order numbers are not the key signal; the real turning point for the industry lies in the "Scaling Law moment" of the robotic brain, where intelligence improves non-linearly with data volume and model scale, breaking through the bottleneck of scenario generalization [1][3]. Group 2: Challenges to Scaling Law Moment - Two major challenges need to be addressed: high hardware costs and the lack of standardized solutions. For instance, Tesla's Optimus Gen1 has a high BOM cost, with a target to reduce it to $20,000 per unit. Key components for cost reduction include joint modules and sensors [3]. - The software side lacks a "robotic version of ChatGPT." The robotic brain must possess both "perception decision-making" and "motion control" capabilities, but current models face data challenges, including complex motion data modalities and high costs of real-world data collection [3][4]. Group 3: Technological Pathways - The "big and small brain collaboration" has become the mainstream engineering approach, with three clear paths for the evolution of large models in robotics. The dual-system layered VLA architecture is currently the optimal solution for engineering implementation [4][5]. - Figure's Helix system exemplifies this collaboration, utilizing a slow system for understanding natural language and a fast system for real-time control, enabling complex tasks in flexible manufacturing scenarios [7][9]. Group 4: Commercialization Pathways - The commercialization of humanoid robots is expected to follow a "from easy to difficult" path, starting with ToG (research and education), then ToB (industrial manufacturing), and finally ToC (household services). The ToB sector is becoming a critical battleground for breakthroughs [8][9]. - The apparel manufacturing industry is a typical case for ToB implementation, with a significant global workforce and high labor costs, yet low penetration of traditional industrial robots due to the flexibility of materials and rapid style changes [8][9]. Group 5: Investment Trends and Future Outlook - The flow of capital in the industry is shifting from a focus on hardware to software, with significant investments in embodied intelligent large models from companies like Google and NVIDIA. Domestic startups are also gaining traction in this space [11]. - The ultimate goal of the humanoid robot industry is to replicate the "non-linear growth curve" seen in sectors like electric vehicles and smartphones, with the "Scaling Law moment" of the robotic brain being the key trigger for this growth [13].
百度及AI的前途
3 6 Ke· 2025-09-24 10:53
Group 1 - Baidu's search engine is undergoing a significant transformation towards AI integration, referred to internally as "Big Search," marking the largest change in a decade [1] - The AI-driven agent model is expected to assist users in completing tasks beyond traditional keyword searches, indicating a shift in user interaction [1] - Baidu's Wenku and cloud storage services are also expanding, aiming to create a "one-stop AI creation platform" with a dedicated team of 1,200 [1] Group 2 - The article discusses the evolution of the internet ecosystem, highlighting the complexity of user needs and the competitive landscape dominated by major players like BAT and FANG [2] - The historical context of the internet's development is explored, noting the transition from information-centric models to more integrated social and e-commerce platforms [3] Group 3 - The recommendation engine developed by Baidu is based on user behavior data, aiming to enhance targeted advertising through detailed user profiling [5] - The article critiques the current state of content production, suggesting that the focus on quantity over quality has led to a decline in meaningful engagement [6] Group 4 - The dominance of algorithm-driven content distribution is noted, with implications for user experience and the overall information ecosystem [8] - Baidu's market position is analyzed in light of competition from ByteDance, emphasizing the challenges faced by traditional search models in adapting to new content consumption patterns [8] Group 5 - The article reflects on the missed opportunities for Baidu in the early days of algorithm distribution, suggesting that a more proactive approach could have altered its competitive stance [11] - The potential of AI to revolutionize information access and user interaction is highlighted, with a focus on the implications for Baidu's future strategies [19][20] Group 6 - Baidu's early commitment to AI, including the establishment of a deep learning research institute, is acknowledged, though recent performance in AI competitions has raised questions about its strategic direction [20] - The article emphasizes the importance of application development in AI, suggesting that successful models will depend on practical use cases rather than theoretical frameworks [32]
在「外滩大会·具身智能:从泛化到行动,重塑产业未来」上,这些大牛都说了什么?
机器之心· 2025-09-16 08:37
Core Viewpoint - The article discusses the future of AI and embodied intelligence, emphasizing the need for disruptive innovation to enable generalized action capabilities and the transition from technical feasibility to commercial success [2][4]. Group 1: Embodied Intelligence Development - The concept of embodied intelligence has evolved from simply giving machines a physical body to creating immersive perception processes [6]. - Current challenges in the field include data bottlenecks, which can be addressed through the establishment of training grounds that enhance robustness and generalization capabilities [7]. - The industry is witnessing a surge in the construction of training grounds, which offer benefits such as cost reduction, safety simulation, and unified standards [7]. Group 2: Data Collection and Utilization - Training grounds are described as new data factories in the AI era, crucial for collecting data to train embodied intelligence models [8][10]. - The development paradigm has shifted to a model where data collection occurs post-robot development, emphasizing the importance of large datasets for effective training [10][11]. - The use of synthetic data is highlighted as a viable solution to the challenges of obtaining real-world data, allowing for scalable and controllable training processes [18][19]. Group 3: Future Prospects and Challenges - The industry is exploring various paths for embodied intelligence, including the integration of real-world data and simulation data to enhance model performance [30][31]. - Discussions on the potential of humanoid robots reveal that while they may not be the only form of embodied intelligence, their development is crucial for achieving broader applications [34][35]. - The timeline for the integration of embodied intelligence into daily life is projected to be gradual, with significant advancements expected in the next 5 to 10 years [38]. Group 4: Industry Collaboration and Ecosystem - The need for collaboration across the industry is emphasized, with calls for the establishment of a robust ecosystem to support the development of embodied intelligence [48][49]. - Various stakeholders express the importance of integrating hardware and software capabilities to enhance the overall effectiveness of embodied intelligence solutions [47][49]. - The article concludes with a vision for a future where embodied intelligence significantly transforms industries and daily life, driven by collective efforts from academia and industry [51].
谁说Scaling Law到头了?新研究:每一步的微小提升会带来指数级增长
3 6 Ke· 2025-09-16 07:46
Core Insights - The Scaling Law is being questioned due to perceived diminishing returns in model training, but recent research suggests that small improvements in accuracy can lead to exponential growth in task completion length, which may hold more economic value in real-world applications [1][2][4] Group 1: Research Findings - A recent paper from Cambridge University indicates that while there are diminishing returns in metrics like test loss, the real-world value of large language models (LLMs) often comes from their ability to complete longer tasks [2][4] - The paper highlights that the long-term execution of tasks has been a significant weakness in deep learning, with LLMs struggling to perform complex, lengthy tasks despite improvements in reasoning capabilities [4][6] - The authors propose that the failures in long tasks are primarily due to execution challenges rather than reasoning or planning limitations, emphasizing the need for more focus on execution capabilities in LLM research [6][20] Group 2: Experimental Insights - The study measures LLMs' long-horizon execution capabilities by isolating execution from planning and knowledge retrieval, revealing that larger models can significantly increase the number of successful execution rounds [6][23][25] - The concept of self-conditioning is introduced, where the model's performance deteriorates as it builds on its previous errors, leading to a decline in accuracy over multiple rounds [8][26][30] - The research shows that while increasing model size improves task execution, it does not alleviate the self-conditioning effect, which remains a challenge for LLMs in long-term tasks [27][30] Group 3: Implications for Investment - The findings suggest that the economic value of LLMs may not be accurately reflected in short-task benchmarks, as the ability to complete longer tasks is a more reliable indicator of their potential [18][20] - The paper encourages further investment in scaling models, as the ability to perform longer tasks could justify continued financial commitment despite short-term performance metrics suggesting stagnation [10][18] - The research calls for the design of new benchmarks that better assess the execution depth of models, highlighting a potential area for future investment and development in the AI sector [10][18]
马斯克周末血裁xAI 500人
Sou Hu Cai Jing· 2025-09-16 06:27
Core Insights - xAI has implemented a sudden internal assessment leading to a significant layoff of its data annotation team, with a reported attrition rate of 33% and over 500 employees terminated [1][11]. Group 1: Layoff Details - The data annotation team, crucial for the development of Grok, has seen its size decrease from 1500 to just over 1000 employees, indicating a nearly one-third reduction [11]. - The layoffs were preceded by a series of one-on-one discussions with employees, creating a sense of panic within the company [5][7]. - The company announced a strategic shift towards hiring specialized data annotators, planning to expand their numbers tenfold, while reducing the focus on general data annotators [11][12]. Group 2: Strategic Shift - This shift from general to specialized data annotation reflects a belief that quality is more important than quantity, aiming to enhance Grok's capabilities in specific fields [12][14]. - The decision may limit the diversity of data available for training, which is essential for the growth of AI systems [12][14]. - The move is seen as a significant gamble on vertical industry AI applications, potentially positioning Grok advantageously if successful [14][15]. Group 3: Management Philosophy - Elon Musk's management style is characterized by a preference for small, high-performing teams, often leading to drastic layoffs to maintain efficiency and performance [22][24]. - This approach has been consistent across Musk's ventures, including Tesla and Twitter, where he has previously enacted similar layoffs to streamline operations [20][24]. - The emphasis on high performance and low tolerance for underachievement is a hallmark of Musk's leadership, which may drive the remaining employees to maximize their potential [22][25].
马斯克周末血裁xAI 500人
量子位· 2025-09-16 05:58
Jay 发自 凹非寺 量子位 | 公众号 QbitAI 什么情况,帮马斯克训练大模型的人说失业就失业了? 马斯克裁员式考核 数据标注团队曾是xAI最大的团队,在Grok的开发过程中发挥了关键作用。他们的工作是标记、分类并将原始数据置于特定语境中,从而教 会AI如何更好地理解世界。 自xAI成立以来,数据标注团队的规模一直在持续增长。 与大多数人工智能公司不同,xAI的许多数据标注员都是直接聘请的,而非外包 。通过这种方式,可以让公司对模型训练拥有更多的控制 权,更好的隐私。 但相应的,成本也更高。 今年2月份,xAI披露计划雇用数千人来帮助训练Grok,并在半年内新增了约700名数据标注员。 上周四晚,xAI内部上演了一场突袭测试,还要求员工必须在第二天早上之前完成并提交。 这可不是一次简单的随堂测试—— 截至目前,本次xAI内部测试的淘汰率高达33%,已有 超过500名员工 被通知卷铺盖走人。 然而9月初,Linkedin页面显示,负责管理数据标注团队的十几名经理中, 至少已有9位被解雇 。 这次不太寻常的人事变动,为即将到来的剧烈动荡埋下了种子。 之后一段时间内,xAI开始与数据标注团队的部分员工开展 一 ...