Core Viewpoint - The article discusses the launch of the new flagship model GLM-4.5 by Zhipu AI, which is designed for intelligent agent applications and has been released on HuggingFace and ModelScope platforms [2][3]. Group 1: Model Architecture and Performance - GLM-4.5 utilizes a mixture of experts (MoE) architecture with a total parameter count of 355 billion and 32 billion active parameters, while GLM-4.5-Air has 106 billion total parameters and 12 billion active parameters [4][6]. - The model integrates reasoning, coding, and intelligent agent capabilities, achieving a comprehensive performance ranking in the global top three, and is the leading domestic and open-source model [3][4]. - In comparative tests against models like Claude Code and Kimi-K2, GLM-4.5 demonstrated superior task completion and tool reliability, although it slightly lagged behind Claude-4-Sonnet in some dimensions [8]. Group 2: Cost and Efficiency - The API call pricing for GLM-4.5 is set at 0.8 yuan per million tokens for input and 2 yuan per million tokens for output, making it a cost-effective option [10]. - The high-speed version of the model supports a generation rate of up to 100 tokens per second, catering to high concurrency deployment needs [12]. Group 3: Training Data and Fine-tuning - The training data for GLM-4.5 encompasses 15 trillion tokens of general corpus, supplemented by 8 trillion tokens specifically fine-tuned for coding, reasoning, and agent tasks, enhanced through reinforcement learning [7]. Group 4: Agent Capabilities and Demonstrations - Zhipu AI has released multiple real-world scenario demos to showcase the agent capabilities of GLM-4.5, including a simulated search engine, a video platform simulator, a playable Flappy Bird game, and an automated PPT tool [14].
专为智能体应用打造,智谱新一代旗舰模型GLM-4.5来了!
硬AI·2025-07-29 15:50