Core Viewpoint - The article discusses the emergence of the LLaDA2.1 model from Ant Group, which has achieved a remarkable speed of 892 tokens per second in complex programming tasks, marking a significant advancement over traditional autoregressive models [1][3][11]. Group 1: Model Performance and Features - LLaDA2.1 operates on a 100 billion parameter scale and has transitioned from a research model to a practical tool, demonstrating superior efficiency [3][4]. - The model introduces a dual-mode decoding strategy, allowing users to switch between Speedy Mode and Quality Mode with a single configuration, thus enhancing usability [9][10]. - In Speedy Mode, LLaDA2.1 achieves a peak speed of 892 tokens per second on the HumanEval+ benchmark, while in Quality Mode, it surpasses previous models in various reasoning tasks [11][31]. Group 2: Technical Innovations - The model employs an Error-Correcting Editable (ECE) mechanism, enabling it to generate drafts quickly and then refine them, addressing the limitations of traditional diffusion models [16][21]. - LLaDA2.1 successfully implements reinforcement learning (RL) on a 100 billion scale, enhancing its performance in instruction-following tasks and demonstrating that diffusion models can achieve both speed and understanding [23][26]. - The introduction of the EBPO algorithm allows for efficient training and editing, marking a significant milestone in the application of RL to diffusion models [25][28]. Group 3: Competitive Advantage - LLaDA2.1's performance in benchmark tests shows a significant advantage over mainstream autoregressive architectures, achieving high speeds without compromising quality [29][30]. - The model's ability to maintain quality even in Speedy Mode demonstrates its robustness, achieving a balance between speed and accuracy [32]. - A lighter 16 billion parameter Mini version has been released, achieving peak speeds exceeding 1500 tokens per second, indicating potential for more lightweight deployments [33].
小众架构赢麻了!通过编辑功能让100B扩散模型飙出892 tokens/秒的速度!
量子位·2026-02-11 01:55