Core Viewpoint - Huawei's Pangu Pro MoE model has been recognized for its innovative design that utilizes dynamic activation of expert networks, achieving superior performance. However, a recent GitHub study claims that the model shares a "striking similarity" in parameter structure with Alibaba's Qwen-2.5 14B model [1]. Group 1: Model Development and Innovation - The Pangu Pro MoE model is developed and trained on the Ascend hardware platform and is not based on incremental training from other vendors' models. It features significant innovations in architecture and technical characteristics [2]. - The model introduces the Grouped Mixture of Experts (MoGE) architecture, which effectively addresses load balancing challenges in large-scale distributed training, thereby enhancing training efficiency [1][2]. Group 2: Open Source Compliance and Community Engagement - Huawei emphasizes that some foundational components of the Pangu Pro MoE model's code implementation reference industry open-source practices and include portions of open-source code from other models. The company adheres strictly to open-source license requirements and clearly marks copyright statements in the open-source code files [2]. - The company promotes an open innovation approach, respecting third-party intellectual property, and advocates for an inclusive, fair, open, united, and sustainable open-source philosophy [2]. - Huawei expresses gratitude to global developers and partners for their support of the Pangu model and highlights the importance of constructive feedback from the open-source community [2].
华为盘古团队声明:严格遵循开源要求
Guan Cha Zhe Wang·2025-07-05 09:32