昇思MindSpore
Search documents
国产AI登顶全球!智谱+华为联手
Ke Ji Ri Bao· 2026-01-17 00:19
Core Insights - GLM-Image, a multimodal image generation model jointly developed by Zhipu and Huawei, has topped the Trending chart on Hugging Face, breaking the long-standing dominance of foreign models in the open-source space [2] - The model is the first state-of-the-art (SOTA) multimodal model trained entirely on domestic chips, showcasing a significant breakthrough in the domestic AI industry chain [2][5] Group 1: Model Architecture and Performance - GLM-Image employs a self-innovated "autoregressive + diffusion decoder" hybrid architecture, enabling the integration of image generation and language models, marking an important exploration in the new generation of "cognitive generation" technology [3] - The model excels in generating text-heavy content, achieving the top rank in the CVTG-2K and LongText-Bench benchmarks, demonstrating superior accuracy in generating multiple text areas within images and rendering long texts [3][6] Group 2: Cost and Efficiency - The model offers high cost-effectiveness, with the API call cost for generating an image being only 0.1 yuan, and a speed-optimized version is set to be released soon [4] Group 3: Domestic Chip Utilization - GLM-Image represents a deep exploration and validation of the domestic computing ecosystem, with all processes from data preprocessing to large-scale pre-training conducted on Huawei's Ascend Atlas 800T A2 devices [5] - This model's development on domestic hardware and frameworks addresses the critical issue of dependency on foreign chips, validating the feasibility of training cutting-edge models on a fully domestic computing stack [5][6] Group 4: Industry Implications - The success of GLM-Image is seen as a result of the collaborative capabilities of the domestic AI industry chain, which can enable small and medium enterprises in China to access AI tools at lower costs and promote domestic AI technology on a global scale [6]
首次!国芯训国模取得世界第一
智通财经网· 2026-01-16 00:33
Core Viewpoint - The collaboration between Zhiyu (02513) and Huawei has led to the development of the GLM-Image model, which is the first state-of-the-art (SOTA) multimodal model trained entirely on domestic chips, marking a significant breakthrough in China's AI model development on the international stage [1][3]. Group 1: Model Development and Performance - GLM-Image was trained using Huawei's Ascend Atlas 800T A2 devices and the MindSpore AI framework, achieving full-process training and inference adaptation [5]. - The model reached the top position on the Hugging Face global AI open-source community leaderboard within 24 hours of its release, indicating its SOTA performance and innovative structure [1][3]. - GLM-Image employs a novel "autoregressive + diffusion decoder" hybrid architecture, which excels in generating knowledge-intensive scenarios such as posters and educational graphics, particularly in generating Chinese characters [4]. Group 2: Technological Significance - This model represents the first fully domestically trained AI model, showcasing China's independent research and development capabilities in AI on an international level [3]. - The collaboration highlights a complete domestic AI technology stack, with Zhiyu's leading model architecture, Huawei's high-performance AI chips, and the self-developed AI computing framework MindSpore, demonstrating a comprehensive breakthrough in core model, hardware, and computing framework [5].
华为公布昇腾AI生态进展:开发者数量400万+,发展合作伙伴3000+
Xin Lang Cai Jing· 2026-01-09 02:34
Core Insights - Huawei has announced significant progress in its Ascend AI ecosystem, targeting developments by the end of 2025, including over 35,000 open-source projects, 400 million developer code contributions, and a developer community of 4 million [1][2][3] Group 1: Open Source and Community Development - The Ascend hardware enabling layer CANN has been fully open-sourced as of December 31, 2025, with an architecture upgrade and over 30 community projects available on the AtomGit platform, attracting more than 1,500 internal and external developers [3][5] - The AI framework MindSpore has achieved over 13 million global downloads, covering more than 130 countries and regions, and has incubated over 40 large models and 380 applications [3][5] Group 2: Industry Solutions and Partnerships - Solutions such as super nodes, large EPs, and integrated machines have been implemented across more than 20 industries, aiding over 2,000 customers in achieving commercial success [3][5] - Huawei has established partnerships with over 3,000 development partners and 80 hardware partners, resulting in the launch of more than 200 products [1][2] Group 3: Talent Development and Competitions - The Ascend AI innovation competition in 2025 attracted 6,310 teams and over 13,600 participants, demonstrating a strong interest in AI talent development [3][5] - Collaborations with 15 top universities have led to the establishment of the Kunpeng Ascend Science and Education Innovation Excellence Center and incubation centers [3][5]
下载量超 1300 万,昇思 MindSpore:AI 框架迈入“超节点时代”
AI前线· 2025-12-30 05:32
Core Insights - The MindSpore community has achieved significant growth, with over 13 million cumulative downloads, more than 52,000 core contributors, and over 120,000 code contributions, serving users in over 150 countries and regions [2] - MindSpore has developed three core capabilities in AI frameworks, focusing on collaboration with training acceleration libraries, model communities, and evaluation tools [3] - The rise of large language models has shifted computational paradigms from single-machine to cluster-based approaches, leading to the development of various parallelization techniques [4] Group 1 - MindSpore supports over 25 model types, providing a comprehensive out-of-the-box capability for script development, parallel training, fine-tuning, and deployment [3] - The framework has achieved over 15% performance improvement in large model inference scenarios through seamless integration with the vLLM community [3] - MindSpore's HyperParallel architecture treats supernodes as a single supercomputer, enhancing programming and scheduling capabilities [6] Group 2 - The HyperParallel architecture introduces key technologies such as Hyperoffload, which separates computation and state to alleviate storage bottlenecks, improving training performance by approximately 20% and increasing sequence length support by about 70% in inference scenarios [4] - MindSpore's native support for ultra-large-scale cluster parallelism can cover tens of thousands of computing nodes and support trillion-parameter models [5] - The framework has been deployed across a wide range of devices, from data center servers to small terminals, establishing itself as a foundational AI capability for numerous smart devices [5] Group 3 - The official version of the HyperParallel architecture and associated acceleration suites for multimodal and reinforcement learning will be released in the first half of next year [7] - Future developments in the MindSpore community will focus on edge intelligence, open architecture, and industry enablement, covering large models and agent acceleration [7] - The introduction of HyperMPMD and Hypershard aims to enhance resource utilization and reduce parallelization modification time significantly [11]
昇思人工智能框架峰会于杭州召开,正式发布“超节点时代”AI框架新范式
Huan Qiu Wang· 2025-12-28 07:13
Core Insights - The summit focused on the "HyperParallel" architecture of the MindSpore AI framework, which aims to meet the increasing demands of large models in terms of computing power, storage, and scheduling efficiency [2][4] - MindSpore has become a leading AI open-source community in China, with over 13 million downloads and contributions from more than 52,000 community members [4] Group 1: HyperParallel Architecture - The HyperParallel architecture introduces three core technologies: HyperOffload, HyperMPMD, and HyperShard, enhancing training performance by over 20% and inference sequence length by 70% [4] - HyperMPMD improves computing resource utilization by over 15% and adapts to complex scenarios like reinforcement learning [4] - HyperShard reduces the time for parallel algorithm adaptation to within one day, significantly increasing tuning efficiency from days to hours [4] Group 2: Industry Applications - In the "AI for Science" sector, MindSpore supports the development of intelligent design systems, such as the "Yufeng·Zhiying" for aerodynamic design, which accelerates traditional processes to real-time interaction [5] - In finance, the application of MindSpore has enabled the stable training of large models with billions of parameters, enhancing service efficiency across various scenarios [7] Group 3: Community and Ecosystem - MindSpore promotes a collaborative open-source philosophy, supporting deployment across various platforms and integrating with mainstream ecosystems [8] - The community has established a talent cultivation system in partnership with educational institutions, training over 400 teachers and covering more than 100 universities [8] Group 4: Future Outlook - The company aims to continue developing an AI framework that is friendly to super nodes, integrates seamlessly across scenarios, and is open and agile, facilitating the intelligent transformation of various industries [9][10]
AI框架迈入超节点时代 国产技术加快产业落地
Xin Lang Cai Jing· 2025-12-26 12:55
Group 1 - The core viewpoint of the articles highlights the transition of AI infrastructure into the "super node era," where traditional server clusters evolve into highly integrated computing systems that support large model training and inference [1][2] - Super nodes are defined as a deep integration of multiple physical machines through high-speed interconnection technology, forming a "supercomputer" that provides resource pooling, scalability, and reliability [1] - The importance of AI frameworks is emphasized as they serve as the core hub connecting computing power and applications, facing unprecedented challenges in parallel scheduling, storage optimization, and programming usability [1] Group 2 - The AI framework adapted to super nodes has demonstrated practical value in key areas, such as the "Yufeng·Zhiyu" intelligent system developed by COMAC, which significantly reduces the simulation cycle for aircraft design from weeks to minutes [2] - In the financial sector, China Merchants Bank has utilized the Ascend framework to create a specialized model with billions of parameters, achieving efficient implementation in scenarios like compliance and customer complaint handling, with a stable training cycle of 1-2 months [2] - The ecological collaboration is identified as a core support for the development of AI frameworks in the super node era, with the MindSpore open-source community gathering 52,000 core contributors and achieving over 13 million downloads, supporting 25 types of mainstream large models and over 3,100 industry applications [2]
昇思MindSpore开源五年下载量超1300万,AI框架进入“超节点时代”
Xin Lang Cai Jing· 2025-12-25 12:14
Core Insights - The conference focused on the theme "MindSpore for Super Nodes," highlighting the innovation in super node technology and the introduction of the HyperParallel architecture to accelerate new model structures and training paradigms in AI frameworks [2][3] Group 1: MindSpore Framework - MindSpore aims to create an AI framework that is super node-friendly, fully integrated across various scenarios, open in architecture, and agile in enabling technology [2] - Since its open-source launch on March 28, 2020, MindSpore has seen over 13 million downloads, covering 156 countries and regions, with more than 120,000 merge requests and over 52,000 community contributors [2] - MindSpore supports over 25 large model series, has 2000+ community partners, and has facilitated over 3100 industry application practices, contributing to nearly 2500 academic papers, ranking first in China and second globally among AI frameworks [2] Group 2: Super Node Technology - The rapid development of AI large model technology is leading to models with long sequences and sparse structures, transitioning AI infrastructure from the "server cluster era" to the "super node era" [3] - The HyperParallel architecture treats super nodes as a "supercomputer" for programming and scheduling, leveraging its advantages to achieve features like HyperShard declarative parallel programming and HyperMPMD heterogeneous irregular parallelism [3] - The framework enhances resource utilization, which is crucial for training large models and practical AI applications, improving task scheduling efficiency compared to other AI frameworks [4]
昇腾“淬火金种子”广深专场激活开发者创新血脉
Huan Qiu Wang· 2025-12-04 08:58
Core Insights - The article emphasizes the importance of solidifying AI technology capabilities and addressing industry implementation challenges as key competitive advantages for companies in the AI sector [1][15] - Huawei's "Ascend" initiative aims to enhance partner capabilities through targeted training, focusing on practical AI applications and ecosystem development [1][13] Group 1: Event Overview - The "Ascend 'Forging Golden Seeds'" training event took place from November 27 to 29, attracting 132 AI developers from 38 partners, focusing on comprehensive technical training [1][10] - The training included hands-on sessions covering foundational architecture, model applications, and programming practices [1][10] Group 2: Training Content - The first day focused on AI application development and optimization, introducing the Ascend hardware and software platform, including the MindIE model inference engine and RAG SDK [3][10] - The second day delved into large model training technologies, featuring the MindSpeed distributed training acceleration suite and the MindSpore AI framework [5][6] - The third day concentrated on low-level CANN operator programming, highlighting the Ascend C programming language for high-performance operator development [8][10] Group 3: Training Methodology - The event broke away from traditional one-way training, integrating technical lectures, developer interactions, and collaborative demand creation [10] - Participants praised the clarity of instruction and the practical nature of the training, indicating a successful learning experience [10][15] Group 4: Future Outlook - Huawei aims to continue collaborating with partners and developers to deeply integrate AI technology into industry practices, accelerating smart upgrades and fostering innovation [15]
华为突破制裁的密码,藏在“384超节点”中
虎嗅APP· 2025-06-17 10:55
Core Viewpoint - The article discusses the challenges and strategies in achieving breakthroughs in artificial intelligence (AI) technology, particularly through the development of Huawei's "CloudMatrix 384 Super Node" computing cluster solution, which aims to overcome limitations in single-point technology by leveraging system engineering innovations [1][3]. Group 1: Huawei's Technological Advancements - Huawei's "CloudMatrix 384 Super Node" is built on 384 Ascend chips and can provide up to 300 PFLOPs of dense BF16 computing power, surpassing NVIDIA's B200 NVL 72 platform [3][4]. - The development of the "Super Node" reflects Huawei's foresight in addressing the diminishing returns of Moore's Law and the increasing costs associated with semiconductor advancements [4][9]. - The architecture of the "Super Node" features a fully interconnected high-speed bus system, enhancing communication bandwidth by 15 times and reducing latency significantly [8][9]. Group 2: System Engineering Innovations - Huawei's approach involves a comprehensive system-level redesign to address challenges in large-scale model training, focusing on resource allocation and communication efficiency [5][10]. - The implementation of global memory unified addressing allows for direct memory access across nodes, improving the efficiency of parameter synchronization during model training [8][9]. - The resource scheduling has been upgraded to enable dynamic task distribution based on model structure, optimizing computation and communication time [8][10]. Group 3: Collaborative Ecosystem Development - Huawei has mobilized a large team across various departments to enhance collaboration and innovation in AI infrastructure, showcasing a unique multi-industry cluster advantage [10][12]. - The company emphasizes the importance of ecosystem compatibility, ensuring that its Ascend architecture supports popular deep learning frameworks like PyTorch and TensorFlow [12][13]. - Huawei's commitment to improving the usability of its AI frameworks, such as MindSpore, aims to facilitate a smoother transition for developers accustomed to existing platforms [12][13]. Group 4: Future Prospects and Industry Impact - The advancements in Huawei's computing capabilities are positioned as a significant step for China's AI industry, potentially overcoming technological limitations and fostering innovation [12][13]. - The ongoing development of the Ascend ecosystem is expected to take time, but efforts are being made to enhance compatibility and support for developers [12][13]. - Huawei's recent achievements in large model training, including the Pangu Ultra MoE model, demonstrate the potential of its domestic computing platform to produce world-class AI models [10][12].
从开源共建到生态繁荣:昇思MindSpore支持Day0迁移、一键部署
财联社· 2025-06-12 10:59
Core Viewpoint - The article emphasizes the rapid development of large models and the need for efficient migration and deployment solutions in the AI ecosystem, particularly through the use of MindSpore, which aims to facilitate seamless integration and performance optimization for developers [1][2]. Group 1: Migration Challenges - The first challenge is fast migration, enabling zero-cost migration of third-party framework models while ensuring complete alignment in model accuracy. MindSpore achieves this through a threefold compatibility approach, allowing for zero-code migration of mainstream models and improving training performance by 5% while maintaining distributed parallel strategies [4]. - The second challenge is rapid deployment, automating the entire training-to-inference process to make large model deployment as simple as executing a single command [2]. Group 2: Training and Inference Solutions - MindSpore supports Day 0 migration for training, providing a "no-sense intelligent translation" capability across frameworks. It utilizes tools like MindSpeed/Megatron for seamless PyTorch model migration, achieving near-zero migration loss for popular models [4]. - In inference deployment, the vLLM-MindSpore plugin allows for HuggingFace models to be deployed in under 30 minutes, with an 80% reduction in weight loading time for large models [5][6]. Group 3: Open Source and Community Engagement - Since its open-source inception on March 28, 2020, MindSpore has fostered a vibrant developer community, with over 1.2 million downloads and contributions from more than 46,000 developers across 2400 cities [7]. - The company promotes a collaborative ecosystem through community governance, providing free computing resources and knowledge sharing across 20+ technical special interest groups (SIGs) [8].