Workflow
K2 Thinking
icon
Search documents
K2 Thinking再炸场,杨植麟凌晨回答了21个问题
36氪· 2025-11-12 13:35
Core Insights - The article discusses the recent release of K2 Thinking, a large AI model developed by Kimi, highlighting its significant advancements and the implications for the AI industry [5][14][15]. Group 1: Model Release and Features - K2 Thinking is a model with 1 trillion parameters, utilizing a sparse mixture of experts (MoE) architecture, making it one of the largest open-source models available [14]. - The model has shown impressive performance in various benchmark tests, particularly in reasoning and task execution, outperforming GPT-5 in certain assessments [15][16]. - K2 Thinking's operational cost is significantly lower than that of GPT-5, with a token output price of $2.5 per million tokens, which is one-fourth of GPT-5's cost [16]. Group 2: Development and Training Insights - The Kimi team has adopted an open-source approach, engaging with communities like Reddit and Zhihu to discuss the model and gather feedback [7][8]. - The training of K2 Thinking was conducted under constrained conditions, utilizing H800 GPUs with Infiniband, and the team emphasized maximizing the performance of each GPU [29]. - The training cost of K2 Thinking is not officially quantified, as it includes significant research and experimental components that are difficult to measure [29][34]. Group 3: Market Trends and Competitive Landscape - The release of K2 Thinking, along with other models like GLM-4.6 and MiniMax M2, indicates a trend of accelerated innovation in domestic AI models, particularly in the context of supply chain disruptions [28][30]. - Different companies are adopting varied strategies in model development, with Kimi focusing on maximizing performance and capabilities, while others like MiniMax prioritize cost-effectiveness and stability [32][33]. - The article notes that the open-source model ecosystem in China is gaining traction, with international developers increasingly building applications on these models [33].
K2 Thinking再炸场,杨植麟凌晨回答了21个问题
3 6 Ke· 2025-11-11 10:30
Core Insights - The K2 Thinking model, developed by Kimi, has gained significant attention following its release, showcasing advancements in AI model architecture and performance [1][2][8] - The model features a sparse mixture of experts (MoE) architecture with 1 trillion parameters, making it one of the largest open-source models available [7][8] - K2 Thinking has demonstrated superior performance in various benchmark tests, outperforming competitors like GPT-5 in specific tasks [8][9] Group 1: Model Features and Performance - K2 Thinking is designed to enhance task execution capabilities, focusing on agentic abilities rather than just conversational skills [12][18] - The model's training cost has been a topic of discussion, with the co-founder clarifying that the reported $4.6 million is not an official figure and is difficult to quantify due to the research and experimental components involved [18][24] - K2 Thinking's output cost is significantly lower than that of GPT-5, priced at $2.5 per million tokens, which is one-fourth of GPT-5's cost [8] Group 2: Community Engagement and Feedback - The Kimi team engaged with the developer community through an AMA session on Reddit, receiving numerous questions and positive feedback regarding the model's capabilities and open-source approach [2][10] - Developers expressed a desire for smaller versions of K2 Thinking to be deployed in PC environments or enterprise settings, indicating strong interest in practical applications [2][10] - The community's enthusiasm reflects a growing trend in the domestic AI model landscape, with multiple companies releasing competitive models in a short timeframe [9][18] Group 3: Technical Innovations and Future Directions - K2 Thinking incorporates innovative techniques such as INT4 quantization and a focus on long reasoning chains, allowing it to perform complex tasks with multiple tool calls [12][14][35] - The Kimi team is exploring advancements in other modalities, such as visual understanding, although timelines for these developments may be extended [17] - Future iterations, including K3, are expected to incorporate significant architectural changes and new features, with a focus on enhancing model capabilities [40][43]
资源不到万亿 OpenAI 的 1% ,Kimi 新模型超越 GPT-5
Founder Park· 2025-11-07 12:00
Core Insights - Kimi has launched the K2 Thinking model, its strongest open-source thinking model to date, featuring 1 trillion parameters and advanced capabilities [2][3] - K2 Thinking model surpasses both open-source and closed-source counterparts in various benchmark tests, achieving state-of-the-art (SOTA) performance [3][10] - The model can autonomously perform up to 300 rounds of tool calls and multi-turn reasoning, indicating a significant advancement from the previous K2 model [6][20] Benchmark Performance - K2 Thinking achieved a 44.9% SOTA score in the Humanity's Last Exam (HLE), a new benchmark designed to evaluate large models' capabilities [10][13] - The HLE test set includes 2,500 advanced academic questions across over 100 disciplines, contributed by nearly 1,000 experts from 50 countries [10][13] - Initial flagship model scores were below 20%, but advancements have led to scores exceeding 40% across the board [13] Model Development and Paradigms - Kimi's approach transitioned from a focus on "model as agent" to "model as thinking agent," emphasizing multi-turn interactions and tool usage [6][15] - The K2 Thinking model incorporates a framework that allows for better interaction with the external world, enhancing its reasoning capabilities [15][21] - The model's ability to maintain reasoning continuity through multi-step tool calls is a unique feature not supported by competitors like OpenAI's GPT series and Google's Gemini [21][23] Competitive Landscape - Kimi's valuation is significantly lower than that of major competitors, with estimates at 0.5% of OpenAI's and 2% of Anthropic's valuations [26][28] - Despite limited resources, Kimi has managed to outperform larger models like GPT-5 and Grok-4 using less than 1% of the resources [29][30] - The current landscape suggests a potential shift in the AI competition, with the possibility of Chinese companies gaining an edge over American counterparts [30]