Stable-DiffCoder超越自回归模型!扩散模型在代码生成取得新突破
机器之心·2026-02-05 23:45

Core Insights - The article discusses the launch of Stable-DiffCoder, a new diffusion language model developed by Huazhong University of Science and Technology and ByteDance, which aims to explore whether diffusion training can enhance model capabilities beyond traditional autoregressive (AR) models [1] Group 1: Model Performance - Stable-DiffCoder outperformed its AR counterparts and several strong open-source models like Qwen2.5-Coder and DeepSeek-Coder on multiple mainstream code benchmarks, demonstrating the effectiveness of the diffusion training paradigm as a powerful data augmentation method [1] - In the 8B model category, Stable-DiffCoder achieved a score of 79.3 on HumanEval and 83.6 on MBPP, surpassing many existing models [23][24] Group 2: Training Methodology - The model utilizes a continuous pre-training (CPT) approach with Block Diffusion and various stability optimization strategies to enhance performance [1] - The training process is designed to first compress knowledge using AR methods before transitioning to diffusion techniques, which helps in efficiently learning a diffusion language model [15][16] Group 3: Knowledge Learning Challenges - The article highlights challenges in the diffusion process, such as the introduction of noise and incorrect knowledge mapping, which can hinder effective learning [5][11] - It emphasizes the importance of maintaining a clean sample distribution during training to ensure effective knowledge transfer [11][20] Group 4: Future Implications - The release of Stable-DiffCoder suggests a new path for the evolution of large models, indicating that AR models can be used as efficient knowledge compressors while diffusion methods can act as enhancers to elevate model intelligence [31]