scaling

Search documents
Grok4、KIMIK2发布,算力板块业绩预告亮眼
Shanxi Securities· 2025-07-17 10:43
Investment Rating - The report maintains an "Outperform" rating for the communication industry, indicating an expected performance exceeding the benchmark index by over 10% [1][36]. Core Insights - The communication industry has seen significant advancements with the release of Grok4 and Kimi K2, which are expected to enhance capabilities in various applications such as programming and robotics [3][15]. - The earnings forecasts for major players in the server, optical module, and copper connection sectors are promising, with notable year-on-year growth expected [5][16]. - The ongoing global competition in computing power is shifting from model training to service quality and competitive advantages, suggesting a robust outlook for investments in the sector [7][18]. Summary by Sections Industry Investment Rating - The communication industry is rated as "Outperform," with expectations of exceeding the benchmark index by over 10% [1][36]. Industry Trends - Grok4, launched by xAI, boasts a tenfold improvement in reasoning capabilities compared to its predecessor, with applications in complex task execution and programming [3][14]. - Kimi K2, a new MoE model, has achieved state-of-the-art results in several foundational tests, indicating significant advancements in AI capabilities [4][15]. Earnings Forecasts - Industrial Fulian anticipates a net profit of 11.96-12.16 billion yuan for the first half of 2025, reflecting a year-on-year increase of 36.8%-39.1% [5][16]. - Other companies like Guangxun Technology and Huagong Technology also project substantial profit growth, with increases ranging from 30% to 95% year-on-year [5][16]. Investment Recommendations - The report suggests focusing on both overseas and domestic computing power chains, highlighting companies such as Industrial Fulian and Huagong Technology as key players [8][19]. - The ongoing arms race in computing power is expected to yield numerous investment opportunities in the coming years, particularly in the context of domestic algorithm optimization [17][18]. Market Overview - The overall market showed positive performance during the week of July 7-11, 2025, with notable increases in various indices, including a 2.36% rise in the ChiNext Index [8][19]. - Specific sectors such as equipment manufacturers and IoT led the weekly gains, indicating strong investor interest [8][19].
X @Cointelegraph
Cointelegraph· 2025-07-16 18:00
🔍 Can a Layer-1 scale without sharding or rollups?Most L1s rely on these external solutions, each introducing trade-offs.Waterfall Network takes a different route with horizontal DAG architecture and native parallel execution, removing the need for additional layers.This simplifies the protocol while preserving performance and decentralization at the base layer.[Research Marketing] ...
RL for Autonomous Coding — Aakanksha Chowdhery, Reflection.ai
AI Engineer· 2025-07-16 16:18
Large Language Models Evolution - Scaling laws 表明,增加计算量、数据和参数可以提高 Transformer 模型的性能,并推广到其他领域 [2][3] - 随着模型规模的扩大,性能持续提高,并在中等数学难题的解决率上有所体现,尤其是在提示模型展示思维链时 [5][7] - 通过强化学习和人类反馈,模型能够更好地遵循指令,从而实现聊天机器人等应用 [10][11] Inference Time Optimization - 通过生成多个响应并进行多数投票(自洽性),可以在推理时提高性能 [15] - 顺序修改之前的响应,特别是在可以验证答案的领域(如数学和编程),可以显著提高性能 [16][17] - 在可以验证答案的领域,推理时间计算的扩展可以转化为智能 [19] Reinforcement Learning for Autonomous Coding - 强化学习是下一个扩展前沿,特别是在可以自动验证输出的领域 [24] - 经验时代将通过强化学习构建超级智能系统,尤其是在具有自动验证的领域 [25] - 自动编码是一个扩展强化学习的绝佳领域,因为它具有验证输出的能力 [30][31] Challenges in Scaling Reinforcement Learning - 扩展强化学习比扩展 LLM 更具挑战性,因为它需要多个模型副本以及训练和推理循环 [29] - 在强化学习中,奖励模型的奖励函数设计是一个挑战 [29][30] Reflection's Mission - Reflection 致力于构建超级智能,并以自主编码作为根本问题 [33] - Reflection 团队由在 LLM 和强化学习领域有开创性工作的 35 位先驱组成 [33]
X @Token Terminal 📊
Token Terminal 📊· 2025-07-11 16:24
Fastest growing @ethereum L2s, based on 7d % growth in L1 to L2 deposits: https://t.co/z3kprWZWWq ...
X @Token Terminal 📊
Token Terminal 📊· 2025-07-11 12:24
Layer 2 Scaling Solutions - Ethereum Layer 2 解决方案旨在扩展以太坊的覆盖范围和网络效应 [1] - 以太坊 Layer 1 和 Layer 2 的每日交易量数据可供参考 [1]
「0天复刻Manus」的背后,这名95后技术人坚信:“通用Agent一定存在,Agent也有Scaling Law”| 万有引力
AI科技大本营· 2025-07-11 09:10
Core Viewpoint - The emergence of AI Agents, particularly with the launch of Manus, has sparked a new wave of interest and debate in the AI community regarding the capabilities and future of these technologies [2][4]. Group 1: Development of AI Agents - Manus has demonstrated the potential of AI Agents to automate complex tasks, evolving from mere language models to actionable digital assistants capable of self-repair and debugging [2][4]. - The CAMEL AI community has been working on Agent frameworks for two years, leading to the rapid development of the OWL project, which quickly gained traction in the open-source community [6][8]. - OWL achieved over 10,000 stars on GitHub within ten days of its release, indicating strong community interest and engagement [9][10]. Group 2: Community Engagement and Feedback - The OWL project received extensive feedback from the community, resulting in rapid iterations and improvements based on user input [9][10]. - The initial version of OWL was limited to local IDE usage, but subsequent updates included a Web App to enhance user experience, showcasing the power of community contributions [10][11]. Group 3: Technical Challenges and Innovations - The development of OWL involved significant optimizations, including balancing performance and resource consumption, which were critical for user satisfaction [12][13]. - The introduction of tools like the Browser Tool and Terminal Tool Kit has expanded the capabilities of OWL, allowing Agents to perform automated tasks and install dependencies independently [12][13]. Group 4: Scaling and Future Directions - The concept of "Agent Scaling Law" is being explored, suggesting that the number of Agents could correlate with system capabilities, similar to model parameters in traditional AI [20][21]. - The CAMEL team is investigating the potential for multi-agent systems to outperform single-agent systems in various tasks, with evidence supporting this hypothesis [21][22]. Group 5: Perspectives on General Agents - There is ongoing debate about the feasibility of "general Agents," with some believing in their potential while others view them as an overhyped concept [2][4][33]. - The CAMEL framework is positioned as a versatile multi-agent system, allowing developers to tailor solutions to specific business needs, thus supporting the idea of general Agents [33][34]. Group 6: Industry Trends and Future Outlook - The rise of protocols like MCP and A2A is shaping the landscape for Agent development, with both seen as beneficial for streamlining integration and enhancing functionality [30][35]. - The industry anticipates a significant increase in Agent projects by 2025, with a focus on both general and specialized Agents, indicating a robust future for this technology [34][36].
对话千寻高阳:端到端是具身未来,分层模型只是短期过渡
晚点LatePost· 2025-07-10 12:30
Core Viewpoint - The breakthrough in embodied intelligence will not occur in laboratories but in practical applications, indicating a shift from academic research to entrepreneurial ventures in the field [1][5]. Company Overview - Qianxun Intelligent was founded by Gao Yang, a chief scientist and assistant professor at Tsinghua University, and Han Fengtao, a veteran in the domestic robotics industry, to explore the potential of embodied intelligence [2][3]. - The company recently demonstrated its new Moz1 robot, capable of performing intricate tasks such as organizing office supplies [4][3]. Industry Trends - The development of embodied intelligence is currently at a critical scaling moment, similar to the advancements seen with large models like GPT-4, but it may take an additional four to five years for significant breakthroughs [2][29]. - There is a notable difference in the development of embodied intelligence between China and the U.S., with China having advantages in hardware manufacturing and faster repair times for robots [6][7]. Research and Development - Gao Yang transitioned from autonomous driving to robotics, believing that robotics offers more versatility and challenges compared to specialized applications like self-driving cars [10][12]. - The field of embodied intelligence is experiencing a convergence of ideas, with many previously explored paths being deemed unfeasible, leading to a more focused research agenda [12][13]. Technological Framework - Gao Yang defines the stages of embodied intelligence, with the industry currently approaching Level 2, where robots can perform a limited range of tasks in office settings [17][18]. - The preferred approach in the industry is end-to-end systems, particularly the vision-language-action (VLA) model, which integrates visual, linguistic, and action components into a unified framework [19][20]. Data and Training - The training of VLA models involves extensive data collection from the internet, followed by fine-tuning with real-world operation data and reinforcement learning to enhance performance [23][24]. - The scaling law observed in the field indicates that increasing data volume significantly improves model performance, with a ratio of 10-fold data increase leading to substantial performance gains [27][28]. Market Dynamics - The demand for humanoid robots stems from the need to operate in environments designed for humans, although non-humanoid designs may also be effective depending on the application [33][34]. - The industry is moving towards a model where both the "brain" (AI) and the "body" (robotic hardware) are developed in tandem, similar to the automotive industry, allowing for specialization in various components [39][41].
「Tokens是胡扯」,Mamba作者抛出颠覆性观点,揭露Transformer深层缺陷
机器之心· 2025-07-09 09:52
Core Viewpoint - The article discusses the trade-offs between State Space Models (SSM) and Transformers, arguing that tokenization is a limitation that SSM can overcome, leading to better computational efficiency and modeling capabilities [1][3][61]. Group 1: State Space Models (SSM) - SSM is defined as a modern version of recurrent neural networks (RNN) with key features that allow it to match the language modeling performance of Transformers [8][10]. - A significant characteristic of SSM is that its hidden state dimension is greater than the input and output dimensions, allowing for better context storage [9][10]. - The model's state update function must be expressive enough to accurately encode and retrieve necessary information, which is achieved through dynamic transfer matrices in selective SSM [11][12]. - Mamba, a specific SSM, integrates parallelization and memory management techniques to enhance computational efficiency [13][14]. - The article highlights that SSMs can outperform Transformers in language modeling tasks when computational resources are matched [53][56]. Group 2: Transformers - Transformers excel in tasks requiring fine-grained operations on individual tokens, but they suffer from quadratic complexity, limiting their efficiency [82][86]. - The article argues that Transformers have an inductive bias that affects their modeling capabilities, making them sensitive to the resolution and semantic content of the data [83][85]. - Despite their strengths, Transformers are not the ultimate solution for all modeling tasks, and there is still significant work to be done in the field [89]. Group 3: Tokenization - Tokenization is a critical step in language modeling, but it introduces limitations in understanding language details [39][40]. - The article posits that removing tokenization could lead to better model performance and aligns with the essence of deep learning, which aims to minimize manual feature engineering [44][45]. - The author suggests that without tokenization, models could learn more effective patterns directly from raw data, enhancing their capabilities [46][52].
为什么 AI 搞不定体力活——对话清华大学刘嘉:这才是生物智能最难攻克的“万里长征” | 万有引力
AI科技大本营· 2025-07-09 07:59
Core Viewpoint - The article discusses the evolution of artificial intelligence (AI) and its intersection with brain science, emphasizing the importance of large models and the historical context of AI development, particularly during its "winters" and the lessons learned from past mistakes [5][18][27]. Group 1: Historical Context of AI - AI experienced significant downturns, known as "AI winters," particularly from the late 1990s to the early 2000s, which led to a lack of interest and investment in the field [2][3]. - Key figures in AI, such as Marvin Minsky, expressed skepticism about the future of AI during these downturns, influencing others like Liu Jia to pivot towards brain science instead [3][14]. - The resurgence of AI began around 2016 with breakthroughs like AlphaGo, prompting a renewed interest in the intersection of brain science and AI [3][14]. Group 2: Lessons from AI Development - Liu Jia reflects on his two-decade absence from AI, realizing that significant advancements in neural networks occurred during this time, which he missed [14][15]. - The article highlights the importance of understanding the "first principles" of AI, particularly the necessity of large models for achieving intelligence [22][27]. - Liu Jia emphasizes that the evolution of AI should not only focus on increasing model size but also on enhancing the complexity of neural networks, drawing parallels with biological evolution [24][25]. Group 3: Current Trends and Future Directions - The article discusses the current landscape of AI, where large models dominate, and the importance of scaling laws in AI development [27][30]. - It notes the competitive nature of the AI industry, where advancements can lead to rapid obsolescence of existing models and companies [36][39]. - The article suggests that future AI development should integrate insights from brain science to create more sophisticated neural networks, moving beyond traditional models [25][50].