Workflow
推理服务器
icon
Search documents
8个字引爆国产芯片!UE8M0 FP8到底是个啥?
华尔街见闻· 2025-08-24 12:54
Core Viewpoint - The introduction of the UE8M0 FP8 format and the next-generation domestic chips signifies a major advancement in the Chinese AI industry, indicating a shift towards hardware-software synergy and reduced reliance on foreign computing power [1][36]. Group 1: Market Reaction - Domestic chip companies' stock prices surged, with Cambrian Technology rising nearly 14% and hitting the daily limit, becoming the top company on the STAR Market [2][35]. - The semiconductor ETF also experienced a significant increase of 5.89% within half a day [5]. Group 2: Technical Explanation of UE8M0 FP8 - UE8M0 FP8 is based on the MXFP8 format, which was defined in the Open Compute Project's 2023 specification, allowing for an 8-bit micro-scaling block format [10][11]. - The MXFP8 format allows for block-level scaling, which retains the 8-bit width while expanding the dynamic range significantly [14][22]. - The UE8M0 format specifically uses an all-exponent representation for scaling factors, which simplifies processing and reduces computational overhead [22][23]. Group 3: Implications for Domestic Chips - Most existing domestic AI accelerators still use FP16/BF16 + INT8 pathways, but new chips like MUSA 3.1 GPU and VIP9000 NPU are beginning to support native FP8 [26]. - The UE8M0 format can save 75% of bandwidth compared to traditional FP32 scaling, making it a crucial optimization for next-generation architectures [27]. - The support for UE8M0 FP8 by various domestic chip manufacturers indicates a significant competitive advantage and potential for improved performance in AI applications [36][37]. Group 4: Industry Collaboration - DeepSeek has collaborated with 15 manufacturers to validate the UE8M0 format, with several companies already adapting their products to support this new standard [26][29]. - The collaboration between DeepSeek and domestic chip manufacturers is likened to the historical Wintel alliance, suggesting a strong ecosystem is being built around these technologies [37].
DeepSeek一句话让国产芯片集体暴涨!背后的UE8M0 FP8到底是个啥
量子位· 2025-08-22 05:51
克雷西 一水 发自 凹非寺 量子位 | 公众号 QbitAI DeepSeek V3.1发布后,一则官方留言让整个AI圈都轰动了: 新的架构、下一代国产芯片,总共短短不到20个字,却蕴含了巨大信息量。 国产芯片企业股价也跟风上涨,比如寒武纪今日早盘盘中大涨近14%,总市值跃居科创板头名。 半导体ETF,同样也是在半天的时间里大涨5.89%。 (不知道作为放出消息的DeepSeek背后公司幻方量化,有没有趁机炒一波【手动狗 头】) | Cambricon | 其武红 | + | | | | | | | | | | | 每日Ali "股票i | | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | | | | | | | SH 688256 ■ Level1基础行情 ■ 上海交易所 ■ 沪港通标的股票 ■ 科创板 ■ 融资融券标的 | | | | | | | | | | 所属行业 × 半导体 +2.68% > | | | | | | | | | | | | | | | 1164.45元 +128. ...
对话红帽全球副总裁曹衡康:AI成本下降了 芯片的量一定会起来
Mei Ri Jing Ji Xin Wen· 2025-06-14 09:02
Core Viewpoint - The consensus in the industry is that the cost of computing power will eventually decrease, but there is no unified path chosen among data centers, integrated machines, or inference servers [1] Group 1: AI Inference Year - 2023 is considered the year of AI inference, marking the official launch of AI applications that will generate business revenue and internal cost control for enterprises [1] - Red Hat has chosen to adopt the "vLLM" framework, a high-performance large language model inference framework that has become a de facto standard in the open-source community [1] Group 2: Contribution and Market Potential - Contributors from China account for 35% of the contributions to the vLLM community, indicating a strong potential for inference technology to bring enterprise value in China [1] - The company identifies two technical challenges in inference: achieving high-performance inference with minimal hardware and cost, and distributing inference workloads across multiple servers [1] Group 3: Future of Computing Power Costs - Red Hat plans to launch inference servers in 2025, emphasizing that the main advantage is the reduction of computing power costs for enterprises [2] - The company does not produce hardware but focuses on software solutions, aiming to lower the barriers for AI adoption among businesses [2] - As computing costs decrease, the demand for GPU cards is expected to rise significantly, potentially increasing the number of enterprises using AI from 1,000 to 100,000 or even 1 million [2]