Workflow
端到端智能体强化学习
icon
Search documents
再给老外亿点点震撼?Kimi杨植麟:啥时发K3? 奥特曼的万亿数据中心建成前
Hua Er Jie Jian Wen· 2025-11-12 13:05
Core Insights - The release of the Kimi K2 Thinking model has generated significant excitement in the AI community, outperforming OpenAI's GPT-5 and Anthropic's Claude Sonnet 4.5 in key benchmark tests while offering a lower API call price [1][6]. Development and Cost - The K2 Thinking model's training cost has been a topic of speculation, with a rumored cost of $4.6 million being dismissed by the founders as unofficial and difficult to quantify due to the research and experimental components involved [7][9]. - The model utilizes a mixed expert architecture with 1 trillion parameters, activating only 32 billion parameters during inference, and employs native INT4 quantization to double inference speed [9]. - The API call pricing is set at 1-4 RMB per million tokens for input and 16 RMB for output, making it one-fourth the cost of GPT-5, thus attracting enterprises to switch from closed-source to open-source solutions [9]. Technical Features and Challenges - The K2 Thinking model prioritizes absolute performance over token efficiency, with plans to incorporate efficiency into future iterations [10][11]. - The development team faced challenges in implementing a "thinking-tool-thinking-tool" model, which is a relatively new behavior in large language models (LLMs) [14]. - The model is designed to perform 200-300 tool calls in sequence to solve complex problems, reflecting a focus on quality in task completion [13]. Future Developments - The timeline for the release of K3 remains uncertain, humorously linked to the completion of a data center by Sam Altman [15]. - The team has opted to release a text model first due to the time required for data acquisition and training adjustments for multimodal capabilities [15]. - The founders expressed a commitment to open-source principles, believing that AGI should promote unity rather than division [17][18]. Licensing and Safety - K2 Thinking is released under a Modified MIT License, requiring attribution for commercial products exceeding 1 million monthly active users or $20 million in monthly revenue [18]. - The founders hinted at the possibility of releasing larger closed-source models if safety concerns arise [19]. Popularity and Community Engagement - Within 48 hours of its release, K2 Thinking achieved over 50,000 downloads, becoming the most popular open-source model on Hugging Face [21]. - The team has expressed a preference for focusing on feature space improvements rather than following the OCR route taken by competitors [22].
月之暗面「调教」出最强Agent,在「人类最后一场考试」拿下最新 SOTA
机器之心· 2025-06-21 05:06
Core Viewpoint - Kimi-Researcher is an advanced autonomous agent developed using end-to-end reinforcement learning, showcasing significant improvements in multi-step reasoning and search capabilities, achieving state-of-the-art performance in various benchmarks [2][4][3]. Group 1: Performance Metrics - Kimi-Researcher achieved a Pass@1 score of 26.9% and a Pass@4 accuracy of 40.17% in the "Humanity's Last Exam," marking a substantial improvement from an initial score of 8.6% [3][4]. - In the xbench-DeepSearch subtask, Kimi-Researcher reached an average Pass@1 score of 69%, outperforming other models equipped with search tools [4]. Group 2: Training Methodology - The agent is trained using end-to-end reinforcement learning, which allows it to learn from a single model that integrates planning, perception, and tool usage without the need for manual rule creation [14][24]. - The training process incorporates a reward mechanism based on final outcomes, ensuring consistent preference direction in dynamic environments [24]. Group 3: Context Management and Efficiency - Kimi-Researcher employs a context management mechanism that enables it to retain key information while discarding irrelevant documents, allowing for over 50 iterations in a single trajectory [27][30]. - The model's training efficiency is enhanced through the introduction of a gamma decay factor, encouraging the discovery of shorter and more efficient exploration paths [25]. Group 4: Tool Utilization and Task Design - The training tasks are designed to necessitate the use of specific tools, promoting the agent's learning on when and how to effectively utilize multiple tools in complex environments [21]. - Kimi-Researcher is capable of conducting academic research, legal and policy analysis, clinical evidence review, and corporate financial analysis, showcasing its versatility [11][8]. Group 5: Infrastructure and Scalability - A scalable, asynchronous rollout system has been developed to enhance the efficiency of agent interactions and reward calculations, significantly improving operational performance [34][32]. - The infrastructure supports dynamic resource allocation and fault tolerance, ensuring high availability in production environments [34].