KTransformers系统

Search documents
高性能计算群星闪耀时
雷峰网· 2025-08-18 11:37
Core Viewpoint - The article emphasizes the critical role of high-performance computing (HPC) in the development and optimization of large language models (LLMs), highlighting the synergy between hardware and software in achieving efficient model training and inference [2][4][19]. Group 1: HPC's Role in LLM Development - HPC has become essential for LLMs, with a significant increase in researchers from HPC backgrounds contributing to system software optimization [2][4]. - The evolution of HPC in China has gone through three main stages, from self-developed computers to the current era of supercomputers built with self-developed processors [4][5]. - Tsinghua University's HPC research institute has played a pioneering role in China's HPC development, focusing on software optimization for large-scale cluster systems [5][11]. Group 2: Key Figures in HPC and AI - Zheng Weimin is recognized as a pioneer in China's HPC and storage fields, contributing significantly to the development of scalable storage solutions and cloud computing platforms [5][13]. - The article discusses the transition of Tsinghua's HPC research focus from traditional computing to storage optimization, driven by the increasing importance of data handling in AI applications [12][13]. - Key researchers like Chen Wenguang and Zhai Jidong have shifted their focus to AI systems software, contributing to the development of frameworks for optimizing large models [29][31]. Group 3: Innovations in Model Training and Inference - The article details the development of the "Eight Trigrams Furnace" system for training large models, which significantly improved the efficiency of training processes [37][39]. - Innovations such as FastMoE and SmartMoE frameworks have emerged to optimize the training of mixture of experts (MoE) models, showcasing the ongoing advancements in model training techniques [41][42]. - The Mooncake and KTransformers systems have been developed to enhance inference efficiency for large models, utilizing shared storage to reduce computational costs [55][57].
院士郑纬民:中国不仅要构建类CUDA系统,同时也要做好10个关键软件
Guan Cha Zhe Wang· 2025-07-26 14:48
Group 1 - The "China Electronic Cloud Artificial Intelligence Innovation Development Forum" was held in Shanghai, focusing on the development of the digital intelligence industry and AI innovation applications [1] - Shanghai is seizing strategic opportunities to deepen the layout of the entire AI industry chain, aiming to exceed 100,000 PetaFLOPS in intelligent computing capacity by the end of 2025 [1] - The city is developing a unique route of "4 basic models + N vertical models" and plans to enhance the quality of data supply systems and accelerate the construction of an AI "highland" [1] Group 2 - Challenges in China's AI industry include issues in chips, computing power, data, and ecosystem, with a focus on developing low-cost personal inference machines and improving the usability of domestic intelligent computing systems [3] - The KTransformers system is highlighted as a way to make AI more accessible through a storage-to-computation approach [3] - Companies are encouraged to embrace AI by identifying core issues, utilizing high-quality data, and fine-tuning foundational large models [3] Group 3 - AI is reshaping the world at an unprecedented speed, with high-quality datasets being crucial for training and optimizing large models [5] - The construction of high-quality datasets faces challenges such as unclear objectives, fragmented implementation paths, and weak technical foundations [5] - Strong policy support from national ministries and local governments is driving the development of high-quality datasets, with new data labeling and synthetic data methods providing solutions [5] Group 4 - China Electronics is establishing a complete integrated circuit industry chain and has developed a full-stack innovation base represented by various companies [7] - The CECSTACK cloud platform, developed by China Electronic Cloud, integrates general computing, intelligent computing, and supercomputing to support AI application development [7] - The company aims to inject new momentum into the "AI+" initiative by creating industry models in key sectors such as government, healthcare, and finance [7]