Core Insights - Huawei announced the open-source release of its Pangu Pro MoE model, which includes a 7 billion parameter dense model and a 72 billion parameter mixture of experts model, as a key initiative for building the Ascend ecosystem [1] - A GitHub study revealed a high similarity of 0.927 in attention parameter distribution between Pangu Pro MoE and Alibaba's Qwen-2.5 14B model, significantly exceeding the typical industry variance of below 0.7 [1] - Huawei's Noah's Ark Lab clarified that the Pangu Pro MoE model was developed and trained on its Ascend hardware platform and not based on other vendors' models [4] - An employee from the Pangu team disclosed that there were instances of "shelling," retraining, and watermark removal, indicating unethical practices within the team [5] Summary by Sections Open Source Announcement - Huawei's open-source release of Pangu Pro MoE is seen as a significant step for the Ascend ecosystem [1] - The model includes a 7 billion parameter dense model and a 72 billion parameter mixture of experts model [1] Research Findings - A GitHub analysis indicated a high similarity of 0.927 between Pangu Pro MoE and Alibaba's Qwen-2.5 model, which is well above the industry norm [1] Company Statements - Huawei's Noah's Ark Lab stated that the Pangu Pro MoE model was developed on its own hardware and not through incremental training of other vendors' models [4] - The team acknowledged that some code implementations referenced industry open-source practices [4] Internal Disclosures - An employee revealed unethical practices, including using Qwen 1.5 110B for retraining and efforts to remove watermarks [5] - The employee cited pressure from leadership and internal doubts as factors leading to these practices [5] - The employee left the company due to ethical concerns and decided to expose these practices [5] Current Status - As of now, Huawei has not issued a statement regarding the employee's disclosures [6]
华为盘古否认抄袭阿里后,其大模型员工自曝存在套壳、续训、洗水印
Qi Lu Wan Bao·2025-07-07 03:50