Workflow
千问开源模型
icon
Search documents
阿里巴巴全栈AI布局浮出水面
Zheng Quan Ri Bao Wang· 2026-01-30 09:15
Core Insights - Alibaba's self-developed AI chip "Zhenwu 810E" has officially launched, marking the debut of its PPU (Parallel Processing Unit) and the formation of the "Tongyun Ge" AI ecosystem, which includes Tongyi Lab, Alibaba Cloud, and Pingtouge [1] - The Tongyi Lab has released the Qwen3-Max-Thinking model, boasting over 1 trillion parameters and 36 trillion tokens of pre-training data, making it Alibaba's largest and most capable model to date [1] - The "Zhenwu" PPU has been deployed in multiple clusters on Alibaba Cloud, serving over 400 clients, including major organizations like State Grid and Xpeng Motors [1] Group 1 - The "Zhenwu" PPU utilizes a self-developed parallel computing architecture and inter-chip communication technology, enabling applications in AI training, inference, and autonomous driving [2] - The PPU has been optimized for the Qwen model, providing integrated products and services through Alibaba Cloud's complete AI software stack [2] - The performance and cost-effectiveness of the "Zhenwu" PPU are comparable to Nvidia's H20, showcasing its competitive edge in the market [2] Group 2 - Alibaba has committed 380 billion yuan to AI infrastructure development, with a long-term goal of increasing cloud data center energy capacity tenfold by 2032 [2] - The CFO of Alibaba indicated that the initial investment plan may need to be increased due to the rapid growth in customer orders outpacing server deployment [2] - The launch of the "Zhenwu" PPU reflects years of strategic investment and vertical integration by Alibaba, culminating in a comprehensive AI stack [2] Group 3 - The "Tongyun Ge" model offers significant advantages in efficiency and cost through its fully self-developed approach, breaking down collaboration barriers between chips, cloud, and models [3] - The closed-loop model enhances end-to-end efficiency in AI training and inference by aligning computational needs with targeted optimizations in architecture and deployment [3]