盘古Pro MoE 72B

Search documents
媲美千亿级模型,华为首个开源大模型上线
Xuan Gu Bao· 2025-06-30 23:32
Group 1 - Huawei announced the open-source release of the Pangu model with 70 billion parameters and the Pro MoE 72B model with 720 billion parameters, enhancing the development of large model technology on domestic computing platforms [1] - The Pro MoE 72B model achieves superior performance comparable to trillion-parameter models by dynamically activating a network of experts, with only 160 billion parameters activated during operation [1] - The latest Super CLUE ranking places Huawei's large model as the top domestic model within the trillion-parameter category, indicating significant advancements in the field [1] Group 2 - Huawei's Ascend chips and CANN heterogeneous computing architecture are part of a fully autonomous and optimized closed-loop solution, marking a shift in global AI computing competition towards large-scale system efficiency and ecosystem development [2] - The Ascend super node has been commercially deployed in data centers such as China Telecom, contributing to the growth of the domestic supply chain [2] Group 3 - Huasen Tiancheng has collaborated with Huawei on Ascend chips and AICC intelligent contact center initiatives, indicating a strategic partnership in AI technology [3] - Softcom Power has launched the Softcom Ascend AI workstation to enhance local AI inference and production efficiency, reflecting the industry's push towards localized AI solutions [4]
华为首个!重磅发布!
Zheng Quan Shi Bao· 2025-06-30 04:37
Core Insights - Huawei has announced the open-sourcing of the Pangu 70 billion parameter dense model and the 720 billion parameter mixture of experts model (Pangu Pro MoE 72B), marking a significant step in its Ascend ecosystem strategy to promote AI research and innovation across various industries [1][5] - The Pro MoE 72B model, with 720 billion parameters and 160 billion activated parameters, demonstrates exceptional performance that can rival models with trillion parameters, ranking first among domestic models under the 1 trillion parameter category in the latest Super CLUE rankings [3][4] - Huawei's Pangu models have been successfully implemented in over 30 industries and 500 scenarios, showcasing their value in sectors such as government, finance, manufacturing, healthcare, and more [5] Summary by Sections Open-Sourcing and Model Performance - Huawei's open-sourcing of the Pangu models aims to enhance the development of AI technologies on domestic computing platforms, expanding the Ascend ecosystem [5] - The Pro MoE 72B model's innovative design allows for dynamic activation of expert networks, achieving high performance with fewer activated parameters [3] Technological Advancements - The recent release of the Pangu Ultra MoE model, with a parameter scale of 718 billion, highlights Huawei's advancements in training large-scale models on the Ascend AI computing platform [4] - The Pangu models are built on a fully integrated software and hardware training system, demonstrating Huawei's capability in achieving a self-controlled training process from hardware to software [4] Industry Impact and Strategic Focus - Huawei emphasizes practical applications of its models, focusing on solving real-world problems across various industries rather than merely theoretical advancements [4] - The launch of the Pangu 5.5 model includes five foundational models targeting NLP, multimodal, prediction, scientific computing, and computer vision, positioning them as core drivers for digital transformation in industries [3]
华为首个!重磅发布!
证券时报· 2025-06-30 04:12
Core Viewpoint - Huawei's announcement to open source the Pangu 70 billion parameter dense model and the 720 billion parameter mixture of experts model (Pangu Pro MoE 72B) is a significant step in promoting the development and application of large model technology across various industries, aligning with its Ascend ecosystem strategy [1][7]. Group 1: Model Specifications and Performance - The newly open-sourced Pro MoE 72B model, with 720 billion parameters and 160 billion active parameters, demonstrates exceptional performance that can rival models with over a trillion parameters, according to the latest Super CLUE rankings [3][4]. - Huawei's Pangu Ultra MoE model, launched on May 30, features a parameter scale of 718 billion, showcasing advancements in training performance on the Ascend AI computing platform [4][5]. Group 2: Strategic Implications - The release of these models signifies Huawei's capability to create world-class large models based on its Ascend architecture, achieving a fully controllable training process from hardware to software [5]. - Huawei's unique approach in the large model strategy emphasizes practical applications across various industries, aiming to solve real-world problems and accelerate the intelligent upgrade of numerous sectors [5][7]. Group 3: Industry Impact - The Pangu large models have been implemented in over 30 industries and 500 scenarios, providing significant value in sectors such as government, finance, manufacturing, healthcare, and autonomous driving [5]. - The open-sourcing initiative is expected to attract more developers and vertical industries to create intelligent solutions based on the Pangu models, further enhancing the integration of AI across different fields [7].