Vera Rubin DSX AI工厂参考设计
Search documents
英伟达、Emerald AI联手六大电力巨头共推“算电协同” 打造新一代AI工厂
Zhi Tong Cai Jing· 2026-03-23 13:13
智通财经APP获悉,英伟达(NVDA.US)与Emerald AI正联合美国电力公司AES发电(AES.US)、 Constellation Energy(CEG.US)、Invenergy、新纪元能源(NEE.US)、Nscale Energy & Power以及 Vistra(VST.US),共同推进新一代AI工厂建设。这类工厂不仅能够更快接入电网,还将作为灵活的能源 资产,为电网提供支撑。 据合作方介绍,新一代AI工厂将采用英伟达Vera Rubin DSX AI工厂参考设计,该设计包含DSX Flex软 件库,可实现AI工厂与电网服务的高效对接。 为实现快速部署,AI工厂可在初期利用就近布置的发电与储能设施作为过渡电源,形成混合型AI工厂; 后续再通过灵活调度这些资源为电网供电,从而加快AI工厂的并网进程,同时为更广泛的电力系统提 供支撑。合作方表示,这种模式有助于加快AI算力的上线速度,为客户与社区创造更多价值。 英伟达创始人兼首席执行官黄仁勋表示:"英伟达与Emerald AI正携手构建AI的未来,让性能、效率与 电网响应能力得以即时协同。" 合作企业还指出,Emerald AI的Conducto ...
英伟达(NVDA.US)、Emerald AI联手六大电力巨头共推“算电协同” 打造新一代AI工厂
智通财经网· 2026-03-23 12:44
英伟达创始人兼首席执行官黄仁勋表示:"英伟达与Emerald AI正携手构建AI的未来,让性能、效率与 电网响应能力得以即时协同。" 合作企业还指出,Emerald AI的Conductor平台将统筹算力灵活性,整合现场发电、储能及其他用户侧资 源,实现精准响应电网需求的电力调度。 智通财经APP获悉,英伟达(NVDA.US)与Emerald AI正联合美国电力公司AES发电(AES.US)、 Constellation Energy(CEG.US)、Invenergy、新纪元能源(NEE.US)、Nscale Energy & Power以及 Vistra(VST.US),共同推进新一代AI工厂建设。这类工厂不仅能够更快接入电网,还将作为灵活的能源 资产,为电网提供支撑。 据合作方介绍,新一代AI工厂将采用英伟达Vera Rubin DSX AI工厂参考设计,该设计包含DSX Flex软 件库,可实现AI工厂与电网服务的高效对接。 为实现快速部署,AI工厂可在初期利用就近布置的发电与储能设施作为过渡电源,形成混合型AI工厂; 后续再通过灵活调度这些资源为电网供电,从而加快AI工厂的并网进程,同时为更广泛的电 ...
从GPU到LPU:英伟达大举进攻推理芯片,黄仁勋再落关键一子
Hua Xia Shi Bao· 2026-03-18 00:59
Core Insights - The AI industry is shifting focus from model training to inference, with companies like NVIDIA adapting to this change by introducing new products and strategies [1][3][6] - NVIDIA's CEO Jensen Huang announced the launch of the Groq 3 LPU, a dedicated AI inference chip, during the GTC 2026 event, aiming to capture a significant share of the inference chip market [1][2] - NVIDIA's revenue forecast for its Blackwell and Rubin product lines has doubled to $1 trillion by the end of 2027, indicating strong market confidence [1] Group 1: NVIDIA's Strategic Moves - NVIDIA has launched the Vera Rubin platform, which includes seven new chips, enhancing its capabilities in AI inference [2] - The Groq 3 LPU is designed to significantly increase token throughput from 100 tokens per second to 1500 tokens or more, supporting advanced AI interactions [2] - NVIDIA's acquisition of Groq's core technology assets for approximately $20 billion in December 2025 has positioned the company to leverage Groq's innovations in its product offerings [3] Group 2: Market Trends and Predictions - The market is witnessing a shift in AI chip shipments, with non-GPGPU chips expected to rise from 36% in 2024 to 45% by 2027, while GPGPU shipments will decline from 64% to 55% [3] - The demand for inference capabilities is being driven by the rise of intelligent agents, which focus more on inference rather than training [6] - NVIDIA's introduction of the LPU is a strategic response to the evolving AI compute demands, addressing the need for efficiency and lower latency in inference scenarios [3][6] Group 3: Ecosystem and Infrastructure Development - NVIDIA is enhancing its ecosystem by introducing the NeMoClaw reference architecture, which includes security and privacy features for enterprise AI systems [6] - The company has also launched the Vera Rubin DSX AI Factory reference design, aimed at optimizing AI infrastructure for scalability and performance [6][7] - Huang emphasized that in the AI era, intelligent tokens are the new currency, and AI factories are essential for generating these tokens, highlighting the importance of infrastructure in AI development [7]
「AI新世代」从GPU到LPU:英伟达大举进攻推理芯片,黄仁勋再落关键一子
Hua Xia Shi Bao· 2026-03-17 12:44
Core Insights - NVIDIA is making a significant move into the inference chip market with the introduction of the Groq 3 LPU, as announced by CEO Jensen Huang at GTC 2026 [1] - The AI industry is shifting focus from model training to inference, with NVIDIA aiming to capture this market opportunity [1][6] - By the end of 2027, NVIDIA's Blackwell and Rubin product lines are projected to generate annual revenues of $1 trillion, doubling previous forecasts [1] Group 1: Product Launch and Features - NVIDIA has officially launched the Vera Rubin platform, which includes seven chips, such as Rubin GPU, Vera CPU, and the new Groq 3 LPU, designed to enhance AI inference capabilities [2] - The Groq LPU is expected to increase token throughput from 100 tokens per second to over 1500 tokens, supporting interactive AI agent scenarios [2] - A new rack, Groq LPX, has been introduced to accommodate the Groq accelerator, enhancing decoding performance for AI models [2] Group 2: Market Trends and Strategic Positioning - NVIDIA's interest in the inference chip market has been long-standing, highlighted by its $20 billion acquisition of Groq's core technology assets in December 2025 [3] - The market share of non-GPGPU chips in AI servers is expected to rise from 36% in 2024 to 45% by 2027, while GPGPU chip share will decline from 64% to 55% [3] - The shift in AI computing demand from training to inference is a strategic response by NVIDIA to market changes and competitive pressures [3][8] Group 3: Ecosystem and Infrastructure Development - NVIDIA is addressing the growing demand for inference with new initiatives, including a partnership with OpenAI for specialized inference chips [4] - The company has introduced the Vera Rubin DSX AI Factory reference design, which outlines how to build and operate AI factory infrastructure for optimal performance [7] - NVIDIA's advancements in AI infrastructure aim to maximize productivity and energy efficiency in generating AI tokens [7] Group 4: Competitive Landscape and Future Outlook - The introduction of the LPU does not imply a decline in NVIDIA's GPU business; rather, it is expected to create broader market opportunities through synergy [7] - The ASIC market is becoming increasingly competitive, with several challengers emerging, including Cerebras and Chinese companies like Cambricon and Huawei [8] - The entry of NVIDIA into the inference chip sector is seen as both a challenge and a catalyst for domestic manufacturers, potentially accelerating industry reshuffling and technological upgrades [8]