Workflow
月之暗面Kimi发布MoE架构基础模型K2并同步开源,总参数1T
news flash·2025-07-11 15:00

Core Insights - The company "月之暗面Kimi" has released the MoE architecture foundational model K2, which features a total of 1 trillion parameters and 32 billion active parameters, surpassing other global open-source models in areas such as autonomous programming, tool utilization, and mathematical reasoning [1] Group 1 - K2 utilizes the MuonClip optimizer to achieve efficient training of trillion-parameter models [1] - The model enhances token efficiency to find new pre-training expansion space amid bottlenecks in high-quality data [1] - K2 demonstrates stronger coding capabilities and excels in general agent tasks, showcasing improved capability generalization and practicality across multiple real-world scenarios [1] Group 2 - The new model is currently available for open experience [1]