Workflow
ModelEngine
icon
Search documents
对标英伟达 华为开源AI容器技术Flex:ai 它可使算力平均利用率提升30%
Mei Ri Jing Ji Xin Wen· 2025-11-21 15:08
Core Insights - The rapid development of the AI industry is creating a massive demand for computing power, but the low utilization rate of global computing resources is becoming a significant bottleneck for industry growth [1] - Huawei's new AI container technology, Flex:ai, aims to address the issue of computing resource waste by allowing a single GPU/NPU card to be divided into multiple virtual computing units, improving resource utilization by 30% [1][2] - Flex:ai is positioned to compete with Nvidia's Run:ai, focusing on software innovation to unify management and scheduling of various computing resources without hardware limitations [2] Group 1 - Flex:ai technology can split a single GPU/NPU card into virtual computing units with a precision of 10%, enabling multiple AI workloads to run simultaneously [1] - The technology has been validated in real-world applications, such as the RuiPath model developed in collaboration with Ruijin Hospital, which improved resource utilization from 40% to 70% [3] - Gartner predicts that by 2027, over 75% of AI workloads will be deployed and run using container technology, indicating a shift towards more efficient resource management [3] Group 2 - Flex:ai will be open-sourced in the Magic Engine community, contributing to Huawei's comprehensive ModelEngine open-source ecosystem for AI training and deployment [3] - Unlike Run:ai, which primarily serves the Nvidia GPU ecosystem, Flex:ai supports a broader range of computing resources, including both Nvidia GPUs and Huawei's Ascend NPUs [2]
见证历史!华为 重大发布
Zhong Guo Ji Jin Bao· 2025-11-21 09:49
Core Insights - Huawei has officially launched a groundbreaking AI technology called Flex:ai, which can enhance the average utilization rate of computing resources by 30% in AI workloads where full card computing power is not utilized [1][3] - Flex:ai is a software for pooling and scheduling various types of processors (XPU) based on Kubernetes, and it is set to be open-sourced to the Moqing community, contributing to a complete ModelEngine open-source ecosystem [1][6] - The technology aims to facilitate the industrial application of AI, with Huawei's VP stating that it can unlock the potential of infrastructure and accelerate the democratization of AI [1][6] Group 1: Flex:ai Capabilities - Flex:ai presents three key capabilities: resource slicing, multi-level intelligent scheduling, and cross-node resource aggregation [3][5] - The technology allows for the slicing of a single GPU/NPU card into multiple virtual computing units, achieving a precision of 10%, enabling multiple AI workloads to run simultaneously on a single card [4][5] - Flex:ai's intelligent scheduling can automatically sense cluster loads and resource states, optimizing the allocation of virtualized GPU and NPU resources based on workload priority and requirements [5][6] Group 2: Competitive Advantages - Compared to Nvidia's Run:ai technology, Flex:ai has two unique advantages: virtualization and intelligent scheduling [5][7] - The "remote virtualization" feature allows for the aggregation of idle XPU resources across the cluster into a "shared computing pool," providing ample support for high-demand AI workloads [5][9] - Flex:ai can also ensure that high-priority AI workloads receive the necessary computing resources, even reallocating resources from lower-priority tasks when needed [5][6] Group 3: Ecosystem Development - The release and open-sourcing of Flex:ai will help Huawei build a complete ModelEngine open-source ecosystem, which includes previously released tools like Nexent, AppEngine, DataMate, and UCM [6][9] - The ModelEngine platform offers a one-stop service for data preprocessing, model training, optimization, and deployment, catering to the needs of large model training and inference [6][9] - The open-source nature of Flex:ai encourages collaboration among developers from academia and industry to establish standards for heterogeneous computing virtualization and AI application integration [9]
见证历史!华为,重大发布
Zhong Guo Ji Jin Bao· 2025-11-21 09:42
Core Insights - Huawei has officially launched a groundbreaking AI technology called Flex:ai, which can enhance the average utilization rate of computing resources by 30% in AI workloads where full card computing power is not utilized [1] Group 1: Flex:ai Overview - Flex:ai is a software for pooling and scheduling various types of processors (XPU) built on Kubernetes, similar to Nvidia's Run:ai technology, but with two unique advantages [1] - The technology is positioned as a crucial tool for the industrialization of AI, with Huawei planning to open-source it to the Magic Engine community to build a complete ModelEngine open-source ecosystem [1] Group 2: Key Capabilities of Flex:ai - Flex:ai addresses the need for AI container technology, which is essential in the era of large models, by providing lightweight virtualization that allows seamless migration of model code and runtime environments [2] - The technology can dynamically mount GPU and NPU resources, enhancing overall resource utilization in clusters [2] - It is projected that by 2027, over 75% of AI workloads will be deployed and run using container technology [2] Group 3: Resource Management and Scheduling - Flex:ai can partition a single GPU/NPU card into multiple virtual computing units with a precision of 10%, allowing multiple AI workloads to run simultaneously on a single card [3] - The technology features unique virtualization capabilities, including "remote virtualization," which aggregates idle XPU resources across the cluster into a "shared computing pool" [5] - Flex:ai's intelligent scheduling can automatically sense cluster load and resource status, optimizing resource allocation based on workload priority and requirements [6] Group 4: ModelEngine Open-source Ecosystem - The release and open-sourcing of Flex:ai will support Huawei in building a comprehensive ModelEngine open-source ecosystem, which includes previously released tools like Nexent, AppEngine, DataMate, and UCM [7] - ModelEngine serves as an AI platform for model training, inference, and application development, providing a one-stop service from data preprocessing to model deployment [7] Group 5: Industry Application and Collaboration - Flex:ai is designed to facilitate the practical application of AI container technology across various industries, addressing the diverse needs of AI workloads [9] - The open-source nature of Flex:ai encourages collaboration among academia, industry, and developers to establish standards for heterogeneous computing virtualization and AI application integration [9]
见证历史!华为,重大发布
中国基金报· 2025-11-21 09:33
Core Viewpoint - Huawei has officially launched a groundbreaking AI technology called Flex:ai, which can enhance the average utilization rate of computing resources by 30% in AI workload scenarios where full card computing power is not utilized [2]. Group 1: Flex:ai Overview - Flex:ai is a software for pooling and scheduling various types of processors (XPU) built on Kubernetes, similar to Nvidia's Run:ai technology, but with two unique advantages [2]. - The technology is considered an important tool for the industrialization of AI, and Huawei plans to open-source it to the Magic Engine community to build a complete ModelEngine open-source ecosystem [2][15]. Group 2: Importance of AI Container Technology - In the era of large models, container technology is seen as a natural partner for AI, allowing for the packaging of model code and runtime environments into lightweight images for seamless cross-platform migration [5]. - Container technology can dynamically mount GPU and NPU computing resources, enhancing overall resource utilization in clusters [6]. - It is projected that by 2027, over 75% of AI workloads will be deployed and run using container technology [6]. Group 3: Key Capabilities of Flex:ai - Flex:ai offers three key capabilities: resource slicing, multi-level intelligent scheduling, and cross-node resource aggregation [7]. - The resource slicing feature allows a single GPU/NPU card to be divided into multiple virtual computing units with a precision of up to 10%, enabling multiple AI workloads to run simultaneously on a single card [7]. Group 4: Advantages Over Competitors - Flex:ai has two distinct advantages over Nvidia's Run:ai: virtualization and intelligent scheduling [11]. - The unique "remote virtualization" technology allows for the aggregation of idle XPU computing power across nodes into a shared computing pool without complex distributed task settings [11]. - Flex:ai's intelligent scheduling can automatically sense cluster load and resource status, optimizing the allocation of local and remote virtualized GPU and NPU resources based on workload priorities and requirements [11][12]. Group 5: Contribution to ModelEngine Ecosystem - The release and open-sourcing of Flex:ai will help Huawei build a complete ModelEngine open-source ecosystem, which includes previously released tools like Nexent intelligent framework and AppEngine [15]. - Flex:ai can facilitate the exploration of AI container technology in various industry scenarios, providing foundational capabilities to enhance computing resource utilization [18].
提升中国病理诊断水平,瑞金医院联合华为开源病理大模型
Guan Cha Zhe Wang· 2025-07-06 05:15
Core Viewpoint - The RuiPath pathology model, developed by Ruijin Hospital in collaboration with Huawei, aims to enhance the efficiency and accuracy of pathology diagnostics in China by leveraging AI technology [1][5]. Group 1: Model Development and Features - The RuiPath model is a clinical-grade multimodal pathology model that covers 90% of the annual cancer incidence in China, addressing 19 common cancer types and hundreds of auxiliary diagnostic tasks [1][5]. - The model has achieved state-of-the-art (SOTA) performance in 7 out of 14 auxiliary diagnostic tasks tested against 12 mainstream public datasets, surpassing the performance of Harvard's UNI2 model [4]. - The model's core "visual foundation model" was developed using over one million high-quality digital pathology slides from Ruijin Hospital, utilizing Huawei's AI toolchain for annotation, training, and fine-tuning [2][4]. Group 2: Efficiency and Impact - The implementation of the RuiPath model allows pathologists to increase their daily workload from 200-300 slides to 400-500 or more, significantly improving diagnostic efficiency [5]. - The model aims to standardize digital pathology practices across hospitals in China, enabling easier deployment and reducing training costs for other institutions [5][10]. - The collaboration between Ruijin Hospital and Huawei has streamlined the model training process, allowing for the completion of the RuiPath model development with only a 16-card cluster, making it more accessible for hospitals [10][11]. Group 3: Industry Challenges and Solutions - There is a significant shortage of pathology doctors in China, with only about 20,000 available and a gap of 140,000 needed, highlighting the importance of AI solutions in addressing this challenge [5]. - The partnership has evolved through two phases: digitalization and smart pathology, focusing on data standardization and collaborative model development [7][8]. - The use of Huawei's ModelEngine has transformed the annotation process, allowing pathologists to review over 700 slides in a day, thus enhancing both efficiency and accuracy [10].
瑞金医院与华为开源RuiPath病理模型,为医疗AI发展按下“加速键”
Huan Qiu Wang· 2025-07-03 07:06
Core Insights - The rapid integration of AI technology into the healthcare sector is transforming traditional medical practices, particularly in areas such as imaging, diagnostics, drug development, and health management [1][12] - The Chinese AI+ healthcare market is projected to grow from 31.5 billion yuan in 2023 to over 80 billion yuan by 2025, with a compound annual growth rate of 58.3% [1] - The global AI healthcare market is expected to exceed $1.5 trillion by 2030, with drug development, imaging diagnostics, and health management accounting for over 60% of this growth [1] AI in Pathology - Pathology diagnosis is considered the "gold standard" for disease diagnosis, especially for cancer, but the traditional process is time-consuming and requires high levels of expertise [2] - There is a significant shortage of pathologists in China, leading to unequal distribution of medical resources and increased patient costs [2] RuiPath Model Development - The RuiPath model, developed by Ruijin Hospital in collaboration with Huawei, is a clinical-grade multi-modal pathology model that covers 90% of cancer cases in China and includes over a hundred auxiliary diagnostic tasks [3] - The open-sourced RuiPath model includes a visual foundation model, a multi-cancer test dataset, and a complete practical guide, significantly lowering the barriers for grassroots hospitals to implement AI-assisted diagnostics [4] Global Collaboration and Standardization - Ruijin Hospital has initiated a global multi-center plan to promote the RuiPath model, focusing on improving diagnostic capabilities in resource-poor areas [5] - The collaboration aims to standardize diagnostic results across different centers and enhance the performance and applicability of the RuiPath model [5] Transition to Smart Pathology - The partnership between Huawei and Ruijin Hospital has evolved through clear phases of digitalization and smart pathology, culminating in the development of a large model capable of precise cancer identification [6] - The introduction of Huawei's ModelEngine AI toolchain has transformed the workflow of pathologists, increasing efficiency and reducing the time required for AI application deployment [7][8] Future Implications - The open-sourcing of the RuiPath model is expected to improve pathology diagnosis in grassroots hospitals, alleviating the imbalance in medical resource distribution [12] - Long-term, this collaboration is anticipated to attract more healthcare institutions and tech companies to participate in the development and application of medical AI, enhancing early diagnosis, personalized treatment, and drug development [12][14]