Workflow
智谱发布新一代大模型GLM-4.6 寒武纪、摩尔线程完成适配

Core Insights - The article highlights the significant advancements made by the domestic AI company Zhipu in the development of its new open-source model, GLM-4.6, which showcases enhanced capabilities in coding and other core functionalities [1][3]. Model Performance - GLM-4.6 has achieved a substantial upgrade in code generation capabilities, aligning it with Claude Sonnet 4, making it the strongest coding model in China [1][3]. - The model has demonstrated improvements in long context processing, reasoning ability, information retrieval, text generation, and agent applications, surpassing the performance of DeepSeek-V3.2-Exp [3][4]. - In real-world programming tasks, GLM-4.6 outperformed Claude Sonnet 4 and other domestic models, with over 30% savings in average token consumption compared to GLM-4.5, marking it as the lowest among similar models [4]. Open Source and Ecosystem - GLM-4.6 is positioned as one of the strongest general-purpose open-source models globally, enhancing the competitive stance of domestic large models in the international landscape [3][4]. - The model's testing environment, ClaudeCode, involved 74 real scenario programming tasks, with all test questions and agent trajectories made public for industry verification and reproducibility [4]. Hardware Adaptation - Zhipu announced that GLM-4.6 has been adapted for deployment on Cambricon's leading domestic AI chips, utilizing an FP8+Int4 mixed-precision inference solution, which significantly reduces inference costs while maintaining model accuracy [4][5]. - The adaptation by Moore Threads based on the vLLM inference framework demonstrates the compatibility and rapid adaptation capabilities of the new generation of GPUs [5]. Future Prospects - The collaboration between the GLM series of models and domestic chips is expected to continuously enhance performance and efficiency in both model training and inference, contributing to a more open, controllable, and efficient AI infrastructure [5].