模芯生态创新联盟
Search documents
阶跃星辰发布开源基座模型 Step 3.5 Flash,多家头部芯片厂商已完成适配
Yang Zi Wan Bao Wang· 2026-02-02 05:37
Core Insights - The release of the new open-source Agent base model Step 3.5 Flash aims to enhance real-time Agent workflow scenarios by balancing inference speed, intelligence level, and cost [1] - Step 3.5 Flash achieves a maximum inference speed of 350 tokens per second for single-request code tasks, positioning itself as a preferred choice for Agent applications [1] Group 1: Model Features - Step 3.5 Flash utilizes a sparse MoE architecture, activating approximately 11 billion parameters per token out of a total of 196 billion parameters, significantly improving inference efficiency while maintaining model capability [1] - The model is designed to provide a more efficient and affordable foundational model option for Agent applications [1] Group 2: Industry Collaboration - In July 2025, the company initiated the "MoCore Ecological Innovation Alliance" with nearly 10 chip and infrastructure manufacturers to eliminate technical barriers between chips, models, and platforms [2] - The alliance aims to enhance computing power utilization efficiency through joint optimization, accelerating the application of large models across various industry scenarios [2] - The industry recognizes that deep collaboration between models and computing power will be a crucial pathway for the large-scale application of inference models [2]
超节点火爆 国产AI算力跑出追赶新路线
Zhong Guo Jing Ying Bao· 2025-08-04 07:26
Core Insights - The 2025 World Artificial Intelligence Conference (WAIC) highlighted the significance of "super nodes" in AI computing infrastructure, with Huawei showcasing its Ascend 384 super node, which boasts a computing power of 300 PFLOPs, nearly double that of NVIDIA's GB200 NVL72 system [1][3] - Domestic AI chip manufacturers are increasingly embracing the super node trend, moving beyond mere parameter comparisons to collaborative efforts, as seen in a rare joint appearance of executives from four domestic AI chip companies [2][11] - The demand for AI computing power is rapidly increasing, leading to the emergence of the "super node" concept as a recognized solution to meet the needs of large-scale AI models [3][4] Super Node Development - The super node concept, proposed by NVIDIA, involves connecting multiple high-performance AI servers to form a larger, more powerful computing node, specifically designed for complex AI model calculations [3][4] - Current super node implementations are characterized by high-performance GPUs interconnected within a single node, with a focus on maintaining consistent bandwidth and latency [4][5] - The future of domestic super node solutions will involve maximizing computing power within individual cabinets and connecting multiple cabinets through optical interconnects [6] Industry Collaboration and Innovation - The WAIC showcased various super node products from multiple vendors, including high-density liquid cooling systems and innovative interconnect technologies, indicating a competitive landscape among domestic manufacturers [7][8] - The emergence of the "AI factory" concept by domestic GPU manufacturers aims to address the efficiency bottlenecks in training large models, emphasizing the need for a comprehensive AI training infrastructure [9][10] - The establishment of the "Model-Chip Ecological Innovation Alliance" signifies a deeper integration between domestic AI models and chips, promoting collaboration among various stakeholders in the industry [11][12]
记者观察:大模型行业应集各家所长打通最后一公里
Zheng Quan Shi Bao Wang· 2025-07-29 07:32
Core Insights - The recent World Artificial Intelligence Conference 2025 highlighted a collaborative trend among large model companies, where competitors are supporting each other rather than competing aggressively [1] - The industry is shifting from a pre-training and supervised learning paradigm, pioneered by OpenAI, to a reinforcement learning approach that significantly enhances reasoning capabilities [1] - The key to increasing the penetration rate of large model applications lies in reducing inference costs, as emphasized by industry leaders [1] Group 1 - The establishment of the "Model-Chip Ecological Innovation Alliance" by Jieyue Xingchen, in collaboration with nearly 10 chip manufacturers and computing platforms, aims to enhance model adaptability and computational efficiency [2] - The creation of Shanghai's first AI terminal soft and hard adaptation optimization pilot platform by Wuwen Qinkong focuses on collaborative innovation across various sectors to address common technical challenges [2] - The concept of an AI integrator that combines computing power, algorithms, data, and intelligent agents was proposed by Jieyue Xingchen's co-founder, suggesting a new operational model for the industry [2] Group 2 - The necessity for companies to leverage their unique strengths and collaborate effectively is emphasized as essential for bridging the gap between technological innovation and industrial application in the large model era [3]