Groq LPU单元
Search documents
Feynman架构登场?英伟达GTC大会或首发1.6nm芯片
Hua Er Jie Jian Wen· 2026-02-25 11:40
GTC 2026焦点或从Vera Rubin转向Feynman Chosun Biz的报道指向一个关键信号,英伟达准备在GTC 2026把叙事重心从Vera Rubin转向Feynman。 当前市场高度关注GTC大会。英伟达可能在GTC大会中抛出下一代芯片代号Feynman,并首次公开展 示采用台积电A16,1.6nm工艺的产品方向,这将把市场对其算力路线图的关注点,从Vera Rubin进一步 推向更远的周期。 据Wccftech援引韩国媒体Chosun Biz报道,英伟达的GTC 2026演讲规划已"超越Vera Rubin",今年的大 会可能成为Feynman的首次公开亮相。GTC 2026将于3月15日开幕,活动回到美国加州圣何塞举行。 黄仁勋此前也表示,其主题演讲将展示"从未公开过"的技术。对投资者而言,这类表态往往意味着新一 轮产品节奏与关键供应链选择即将被确认,尤其是先进制程与封装形态的取舍。 如果Feynman确实采用台积电A16,Wccftech认为英伟达将成为该节点初期大规模量产阶段的首家,甚 至可能是唯一客户,这将把先进产能与良率爬坡的市场预期进一步绑定到英伟达。 同时,市场还在评估Fe ...
英伟达封死了ASIC的后路?
半导体行业观察· 2025-12-29 01:53
Core Viewpoint - NVIDIA aims to dominate the inference stack with its next-generation Feynman chip by integrating LPU units into its architecture, leveraging a licensing agreement with Groq for LPU technology [1][18]. Group 1: NVIDIA's Strategy and Technology Integration - NVIDIA plans to integrate Groq's LPU units into its Feynman GPU architecture, potentially using TSMC's hybrid bonding technology for stacking [1][3]. - The LPU modules are expected to enhance inference performance significantly, with Groq's LPU set to debut in 2028 [5]. - The Feynman core will utilize a combination of logic and compute chips, achieving high density and bandwidth while maintaining cost efficiency [6]. Group 2: Inference Market Dynamics - The AI industry's computational demands have shifted towards inference, with major companies like OpenAI and Google focusing on building robust inference stacks [9]. - Google’s Ironwood TPU is positioned as a competitor to NVIDIA, emphasizing the need for low-latency execution engines in large-scale data centers [9][10]. - Groq's LPU architecture is designed specifically for inference workloads, offering deterministic execution and on-chip SRAM for reduced latency [10][14]. Group 3: Licensing Agreement and Market Position - NVIDIA's agreement with Groq is framed as a non-exclusive licensing deal, allowing NVIDIA to integrate Groq's low-latency processors into its AI Factory architecture [18][21]. - This strategy is seen as a way to circumvent antitrust scrutiny while acquiring valuable talent and intellectual property from Groq [19][21]. - The transaction is viewed as a significant achievement for NVIDIA, positioning LPU as a core component of its AI workload strategy [16][21].