真武810E(真武PPU)
Search documents
阿里自研AI芯片“真武”亮相,性能或超英伟达A100
Guan Cha Zhe Wang· 2026-01-29 09:15
Core Insights - Alibaba has launched the "Zhenwu 810E" high-end AI chip, marking the debut of its self-developed PPU, part of the AI supercomputer initiative "Tongyun Ge" [1][3] - The "Tongyun Ge" integrates Alibaba's self-developed chips, leading cloud services, and advanced open-source models to achieve high efficiency in AI model training and deployment [1] - Alibaba and Google are the only two companies globally with top-tier capabilities across large models, cloud, and chip technology [1] Group 1 - The "Zhenwu" PPU features a self-developed parallel computing architecture and inter-chip communication technology, with 96G HBM2e memory and a bandwidth of 700 GB/s, suitable for AI training, inference, and autonomous driving [3] - The "Zhenwu" PPU has been deployed in multiple clusters on Alibaba Cloud, serving over 400 clients, including major organizations like State Grid and Xpeng Motors [1][3] - The performance of the "Zhenwu" PPU surpasses that of Nvidia's A800 and is comparable to the H20, with an upgraded version reportedly outperforming the A100 [3] Group 2 - Tongyi Laboratory released the Qwen3-Max-Thinking model, achieving multiple global records and performance levels comparable to GPT-5.2 and Gemini 3 Pro [4] - The number of derivative models from the Qwen open-source model has exceeded 200,000, with downloads surpassing 1 billion, maintaining its position as the largest globally [4]
性能与英伟达H20相当!阿里自研AI芯片曝光,形成“通云哥”黄金三角
财联社· 2026-01-29 07:52
Core Viewpoint - The launch of the "Zhenwu 810E" high-end AI chip by PingTouGe marks a significant advancement in Alibaba's AI capabilities, showcasing the integration of its AI ecosystem known as "Tongyun Ge" [1][4]. Group 1: Product Launch and Features - The "Zhenwu" PPU has been deployed in multiple large-scale clusters on Alibaba Cloud, serving over 400 clients including State Grid, Chinese Academy of Sciences, Xiaopeng Motors, and Sina Weibo [2][4]. - The "Zhenwu" PPU features a self-developed parallel computing architecture and inter-chip interconnection technology, with a memory of 96G HBM2e and an inter-chip bandwidth of 700 GB/s, suitable for AI training, inference, and autonomous driving [4]. - The performance of the "Zhenwu" PPU surpasses mainstream domestic GPUs and is comparable to NVIDIA's H20, receiving positive feedback for its stability and cost-effectiveness, leading to high market demand [4]. Group 2: Strategic Development and Market Position - The introduction of the "Zhenwu" PPU reflects PingTouGe's years of expertise in the chip sector, with Alibaba investing strategically for 17 years to achieve a comprehensive AI layout through vertical integration [5]. - The Tongyi Laboratory recently released the Qwen3-Max-Thinking flagship inference model, setting multiple global records and performing comparably to GPT-5.2 and Gemini 3 Pro, with over 200,000 derivative models and 1 billion downloads from the Hugging Face community [5]. - Alibaba Cloud operates the leading AI infrastructure in China and is globally recognized, holding over 29 regions with 92 available zones, capturing more market share in the Chinese AI cloud market than the combined total of the second to fourth competitors [5].
阿里证实自研AI芯片PPU,“通云哥”阵型浮出水面
Mei Ri Jing Ji Xin Wen· 2026-01-29 06:29
Core Insights - Alibaba's self-developed AI chip "Zhenwu 810E" has officially launched, marking a significant milestone in its 17-year strategic journey in AI and cloud computing [2][4][6] - The "Zhenwu" PPU chip features a 96GB HBM2e memory and a 700 GB/s interconnect bandwidth, positioning it as a competitive alternative to mainstream domestic GPUs and comparable to NVIDIA's H20 [4][5] - The launch of the "Zhenwu" PPU completes Alibaba's strategic framework of "large models + cloud + chips," establishing it as a leading player in the AI industry [4][7] Product and Technology - The "Zhenwu" PPU utilizes a self-developed parallel computing architecture and inter-chip interconnect technology, along with a fully self-developed software stack [4] - The chip has been in high demand, with multiple deployments in Alibaba Cloud serving over 400 clients, including major organizations like the State Grid and Xpeng Motors [4][7] - The chip's performance has been validated through extensive internal testing, demonstrating its stability and cost-effectiveness [6] Market Position and Strategy - Alibaba Cloud has achieved a 34% revenue growth, with AI-related product revenues increasing for nine consecutive quarters, indicating strong market demand for AI computing resources [7] - The company is prioritizing investments in AI infrastructure to meet the growing demand, with plans to potentially increase capital expenditures [7] - Alibaba's strategy contrasts with competitors like Amazon and Microsoft, focusing on a "full-stack self-research" approach rather than relying on external partnerships [8][9] Competitive Landscape - The global cloud market is dominated by four major players: Amazon, Microsoft, Google, and Alibaba, with Alibaba and Google pursuing a self-research strategy that may offer long-term value [8] - Self-developed chips are seen as a critical factor in reducing total cost of ownership (TCO) for AI computing, with examples like Google's TPU demonstrating significant cost advantages [9] - Alibaba's "Tongyun Ge" framework positions it favorably in the AI competition, enhancing its capabilities across cloud infrastructure, AI models, and chip technology [11]
重磅!阿里,刚刚发布!自研AI芯片来了!
券商中国· 2026-01-29 02:56
Core Viewpoint - Alibaba is establishing an AI supercomputer named "Tongyun Ge," integrating its self-developed chips, cloud services, and advanced models to enhance efficiency in training and deploying large models [2][4]. Group 1: AI Chip Development - The "Zhenwu 810E" high-end AI chip has been launched by Alibaba's Pingtouge, marking the emergence of the "Tongyun Ge" AI triangle consisting of Tongyi Lab, Alibaba Cloud, and Pingtouge [2][4]. - The "Zhenwu" PPU features a self-developed parallel computing architecture with 96G HBM2e memory and an inter-chip bandwidth of 700 GB/s, suitable for AI training, inference, and autonomous driving [8]. Group 2: Model Performance - The Qwen3-Max-Thinking model, released by Tongyi Lab, has set multiple global records, with over 1 trillion parameters and 36 trillion tokens of pre-training data, rivaling models like GPT-5.2 and Gemini 3 Pro [10]. - The model has demonstrated superior performance in 19 benchmark tests, showcasing its capabilities in factual knowledge, complex reasoning, and alignment with human preferences [10]. Group 3: Market Impact and Adoption - The "Zhenwu" PPU has been deployed in over 400 clients, including major organizations like State Grid and Xpeng Motors, indicating strong market demand and positive reception [4][8]. - The Qwen3-Max-Thinking model is available for free on QwenChat, with over 200,000 derivative models and 1 billion downloads, establishing it as a leader in the global AI open-source community [11].