Code World Model (CWM)
Search documents
首个代码世界模型引爆AI圈,能让智能体学会「真推理」,Meta开源
具身智能之心· 2025-09-26 00:04
Core Viewpoint - The article discusses the introduction of the Code World Model (CWM) by Meta, which represents a significant evolution in AI models aimed at improving code generation through world modeling techniques [2][5][31]. Group 1: Model Overview - CWM is a 32 billion parameter open-weight large language model (LLM) designed to enhance code generation research based on world models [7][12]. - It features a dense, decoder-only architecture with a context length of up to 131k tokens, demonstrating strong performance in general programming and mathematical tasks [8][9]. Group 2: Performance Metrics - CWM achieved notable scores in various benchmarks: SWE-bench Verified (pass@1 65.8%), LiveCodeBench (68.6%), Math-500 (96.6%), and AIME 2024 (76.0%) [8][23]. - In comparison to other models, CWM's performance is competitive, particularly in the 30B parameter range [9][30]. Group 3: Training Methodology - The model was trained using extensive observational-action trajectories in a Python interpreter and agent-based Docker environment, focusing on improving code understanding beyond static code training [12][22]. - Meta has made available checkpoints from the mid-training, SFT, and reinforcement learning phases to support further research [13]. Group 4: Research Implications - CWM serves as a robust testing platform to explore the potential of world modeling in enhancing reasoning and planning capabilities in code generation [15][31]. - The research indicates that world models can benefit agent-based coding by allowing for stepwise simulation of Python code execution, which enhances reasoning from such simulations [16][31]. Group 5: Future Directions - Meta envisions that the code world model will bridge the gap between linguistic reasoning and executable semantics, with ongoing research needed to fully leverage its advantages across tasks [31]. - The model aims to improve reinforcement learning by enabling agents familiar with environmental dynamics to focus on learning actions that yield rewards [31].
首个代码世界模型引爆AI圈,能让智能体学会「真推理」,Meta开源
机器之心· 2025-09-25 03:20
Core Insights - The article discusses the introduction of the Code World Model (CWM) by Meta, which is a significant advancement in AI for code generation and reasoning [1][2][4]. Group 1: Model Overview - CWM is a 32 billion parameter open-weight large language model (LLM) designed to enhance code generation through world modeling [7]. - It supports a maximum context length of 131k tokens and is structured as a dense, decoder-only LLM [8]. - The model has shown strong performance in general programming and mathematical tasks, achieving a pass rate of 96.6% on Math-500 and 76.0% on AIME 2024 [6]. Group 2: Training and Methodology - To improve code understanding, the Meta FAIR CodeGen team utilized extensive observation-action trajectories in a Python interpreter and agent-based Docker environment for mid-training [12]. - CWM was trained on a large dataset of coding data and customized Python + Bash world modeling data, enabling it to simulate Python function execution and agent interactions in Bash [22]. Group 3: Performance Metrics - CWM achieved notable performance in various benchmarks, including a pass rate of 35.1% in the Aider Polyglot benchmark and 65.8% in SWE-bench Verified with test-time extension [23][26]. - In comparison to other models, CWM demonstrated competitive results, particularly in time and space complexity predictions, outperforming baseline models in all metrics [29]. Group 4: Future Research Directions - Meta envisions CWM bridging the gap between language-level reasoning and executable semantics, with potential applications in zero-shot planning and reinforcement learning [30]. - The model's ability to predict the consequences of its actions is expected to enhance efficiency in interactions with environments, allowing for more complex task handling [30].