Workflow
算力拐点
icon
Search documents
中金 | GTC 2026:推理时代已至,再绘AI硬件宏伟蓝图
中金点睛· 2026-03-19 00:11
Core Viewpoint - NVIDIA management highlighted that AI inference is entering an "Inference Inflection Point," with industry demand shifting from a "training-driven" to an "inference-driven" paradigm due to the expansion of model size, context length, and token generation [1][6]. Group 1: AI Infrastructure and Hardware Developments - The Vera Rubin platform features a hardware configuration of "72 GPUs + 36 CPUs + NVLink 6 + CX9 + BF4 DPU," designed to meet the increasing demands of inference tasks [4][8]. - NVIDIA's new hardware architecture aims to enhance efficiency, achieving a system-level energy efficiency improvement of 4 times and a maximum inference throughput per watt increase of 10 times compared to previous generations [8]. - The company anticipates that the data center business revenue will surge from $500 billion in 2025-2026 to over $1 trillion in 2026-2027, driven by the exponential growth in AI inference demand [6][7]. Group 2: Market Growth and Projections - The AI PCB market is projected to grow significantly, with an expected market size of $22.464 billion by 2027, reflecting an 86% year-on-year increase [35][45]. - The introduction of Groq LPU and the Rubin architecture is expected to drive a substantial increase in PCB material specifications and usage, contributing to the overall growth of the AI PCB market [35][45]. Group 3: Innovations in Interconnect Technology - NVIDIA's Spectrum-X CPO switch, which integrates optical and electrical components, is now in full production, aiming to reduce power consumption and signal degradation in large-scale computing centers [47][48]. - The trend towards scale-up CPO technology is expected to enhance interconnect capabilities, providing a pathway for high-density, low-power networking solutions in future architectures [48].