Workflow
华为云CloudMatrix 384超节点
icon
Search documents
华为云CloudMatrix 384超节点再出圈,昇腾AI云服务解锁超级算力
Guan Cha Zhe Wang· 2025-07-28 07:15
Core Insights - Huawei's 384 Super Node received the "Treasure of the Museum" award at the World Artificial Intelligence Conference (WAIC 2025), highlighting its significance in the AI landscape [1] - The CloudMatrix 384 Super Node enables flexible and on-demand access to powerful computing resources, making advanced AI infrastructure more accessible for enterprises [2][3] Group 1: Technology Features - The CloudMatrix 384 Super Node integrates 384 Ascend NPUs and 192 Kunpeng CPUs through a new high-speed network, achieving a computing power scale of 300 PFlops and overcoming bandwidth performance bottlenecks [3][4] - It features four key technological advantages: strong throughput performance with 2300 Tokens decoding per card, coverage of over 160 mainstream models for efficient model migration, a pioneering large-scale expert parallel scheme for system-level optimization, and flexible scalability with low initial investment [4] Group 2: Industry Applications - The CloudMatrix 384 Super Node has been widely adopted across various industries, enhancing delivery efficiency by over 50% for Sina's "Smart Xiao Lang," supporting 6 million daily users for Silicon-based Flow, and accelerating AI model training for the Chinese Academy of Sciences [4] - Huawei Cloud's solutions are being utilized in diverse sectors, such as developing a railway model and smart inspection robots in collaboration with North Railway Institute, and enhancing renewable energy generation through AI and meteorology partnerships [10] Group 3: Future Outlook - Huawei Cloud aims to continue leveraging its advanced technologies and industry-specific solutions to address real-world challenges, fostering a new intelligent world in collaboration with clients and partners [10]
封杀中国芯片?!歇斯底里,黔驴技穷!
半导体芯闻· 2025-05-21 10:29
Core Viewpoint - The article discusses the recent aggressive semiconductor export control measures introduced by the U.S. Department of Commerce, which effectively bans the use of Huawei's Ascend 910 series chips globally, labeling it as a violation of U.S. export control regulations, with severe penalties for non-compliance [2][4]. Group 1: U.S. Export Control Measures - The U.S. Department of Commerce has issued guidelines that classify the use of Huawei's Ascend chips as a violation of export controls, threatening legal consequences for any organization or individual that complies with these measures [4][6]. - The measures are described as unilateral bullying and protectionism, which undermine the stability of the global semiconductor supply chain and infringe on the rights of other countries to develop advanced computing and AI technologies [4][8]. Group 2: China's Response - China firmly opposes the U.S. measures, asserting that they violate international law and the basic principles of international relations, and emphasizes the need for cooperation and mutual benefit in technological development [4][8]. - The Chinese government warns that any organization or individual that assists in enforcing U.S. measures may face legal repercussions under Chinese law, specifically referencing the Anti-Foreign Sanctions Law [4][8]. Group 3: Implications for Global Semiconductor Industry - The article highlights the potential disruption to the global semiconductor industry caused by U.S. actions, which could lead to a significant impact on supply chains and technological collaboration [4][9]. - It notes that the U.S. accusations against Huawei's chips are seen as unfounded and part of a broader strategy to maintain technological dominance and suppress China's advancements in the semiconductor sector [9][11].
华为云黄瑾:六大科技创新,昇腾云定义下一代AI基础设施
Jing Ji Wang· 2025-05-19 08:17
Core Viewpoint - The emergence of super-node architecture is a breakthrough in technology and engineering innovation, addressing the explosive growth in computational power demand for AI training and inference, as stated by Huawei Cloud's Vice President Huang Jin [1][3]. Group 1: Technological Innovations - Huawei Cloud's CloudMatrix 384 super-node features six key technological advantages: MoE affinity, network-enhanced computing, memory-enhanced computing, long-term reliability, peak-shifting resource usage, and on-demand availability [1][4]. - The evolution of AI model training has transitioned from small models on single cards to large models on AI clusters, with computational power demand increasing by 10,000 times [3]. - Over the past eight years, single-card hardware performance has increased by 40 times, while inter-node network bandwidth has only increased by 4 times, highlighting the communication bottleneck in AI model training [3]. Group 2: Implementation and Applications - The CloudMatrix super-node clusters have been deployed in Wuhu, Gui'an, and Inner Mongolia, showcasing the practical application of the new architecture [4]. - Huawei Cloud aims to transform AI technology value into practical applications across various industries, emphasizing the importance of AI as a universal technology [5]. - The company has adapted its Ascend AI cloud services to over 160 large models, assisting clients in model development, training, hosting, and application across sectors such as government, finance, retail, internet, transportation, and manufacturing [5].
华为云黄瑾:传统计算架构难支撑AI代际跃迁,超节点架构是创新
Bei Ke Cai Jing· 2025-05-16 12:56
Core Insights - The rapid growth in demand for AI computing power has outpaced the capabilities of traditional computing architectures, necessitating the development of new solutions like the super node architecture [1] - Huawei Cloud's CloudMatrix 384 super node addresses key technical challenges in AI computing, including communication efficiency, memory limitations, and reliability, achieving a computing power scale of up to 300 Pflops, surpassing NVIDIA's NVL72 by 67% [1] - The introduction of distributed inference platforms and innovative technologies such as Elastic Memory Storage (EMS) significantly enhances resource utilization and performance, reducing latency and improving fault detection rates [2] Group 1 - The demand for AI computing power has increased by 10,000 times, while hardware capabilities have only improved by 40 times in the last eight years [1] - The CloudMatrix 384 super node connects 384 cards into a single super cloud server using a new high-speed interconnect bus [1] - The super node features six technical advantages, including MoE affinity and high reliability [1] Group 2 - The distributed inference platform allows for efficient distributed inference with one card acting as one expert, significantly improving MoE computation and communication efficiency [2] - The MatrixLink service consists of two network layers, enabling high-speed interconnection within the super node and low latency communication [2] - The EMS technology decouples memory from computing power, enhancing resource utilization and reducing the first token latency by up to 80% [2]
华为云黄瑾:传统计算架构已难以支撑AI技术代际跃迁,超节点架构开辟AI产业新路径
Xin Lang Ke Ji· 2025-05-16 11:25
Core Viewpoint - Huawei Cloud introduced the CloudMatrix 384 super node technology to address the explosive growth in computing power demand for large model training and inference, marking a significant engineering innovation in the AI industry [1][2]. Group 1: Technology and Innovation - The CloudMatrix 384 super node architecture overcomes traditional computing limitations by addressing three major technical challenges: communication efficiency bottlenecks, memory wall constraints, and reliability shortcomings [2]. - The new architecture utilizes a high-speed interconnect bus to link 384 cards, forming a super cloud server capable of delivering up to 300 Pflops of computing power, which is 67% ahead of similar products in the industry [2]. Group 2: Market Application and Strategy - Huawei Cloud's strategy emphasizes the "Ascend Cloud Service supporting hundreds of models and thousands of scenarios," aiming to enhance AI cloud services for both domestic and global clients [3]. - The Ascend AI cloud service has been adapted for over 160 large models, including DeepSeek, and has served more than 600 innovative enterprises across various sectors such as government, finance, retail, internet, transportation, and manufacturing [3].