Workflow
chiplet设计
icon
Search documents
英伟达下一代GPU,巨幅升级!
半导体芯闻· 2025-09-29 09:45
Core Insights - NVIDIA and AMD are competing to develop superior AI architectures, with significant upgrades planned for their next-generation products in terms of power consumption, memory bandwidth, and process node utilization [1][2] - AMD's Instinct MI450 AI series is expected to be highly competitive against NVIDIA's Vera Rubin, with both companies making substantial modifications to their designs [1][5] Group 1: AMD's Optimism and Product Comparison - AMD executive Forrest Norrod expressed optimism about the MI450 product line, likening it to AMD's transformative "Milan moment" with the EPYC 7003 series [2] - Norrod stated that MI450 will be more competitive than NVIDIA's Vera Rubin and will utilize AMD's technology stack [3] - The MI450X's TGP has increased by 200W, while Rubin's TGP has risen by 500W to 2300W, indicating a significant enhancement in performance [5] Group 2: Specifications and Technological Advancements - The MI450 is rumored to launch in 2026 with HBM4 memory, offering up to 432 GB per GPU and a memory bandwidth of approximately 19.6 TB/s, while Vera Rubin is expected to have around 288 GB per GPU and a bandwidth of ~20 TB/s [6] - AMD's dense compute performance is estimated at ~40 PFLOPS, compared to ~50 PFLOPS for NVIDIA's offering [6] - Both companies are expected to narrow the technological gap as they adopt similar technologies, including HBM4 and TSMC's N3P node [6] Group 3: D2D Interconnect Technology - AMD plans to significantly enhance its D2D interconnect technology with the upcoming Zen 6 processors, as evidenced by developments in the Strix Halo APU [7][8] - The current D2D communication method using SERDES has limitations in efficiency and latency, which AMD aims to address with new designs [10][12] - The Strix Halo utilizes TSMC's InFO-oS and redistribution layer (RDL) to improve communication between chips, reducing power consumption and latency [12][14]
英特尔最新芯片,全用台积电?
半导体芯闻· 2025-05-06 11:08
Core Insights - Intel's Arrow Lake architecture features a chiplet design, showcasing a complex layout of compute, I/O, SoC, and GPU tiles, with filler dies for structural support [1][3] - The compute tile utilizes TSMC's advanced N3B process, while the I/O and SoC tiles are manufactured using the older N6 process, indicating a significant reliance on competitor technology [3] - Arrow Lake introduces a new cache structure, allowing E-core clusters to access shared L3 cache, enhancing performance capabilities [5] Architecture Details - The compute tile measures 117.241 mm², while the I/O and SoC tiles are 24.475 mm² and 86.648 mm² respectively, all mounted on a base tile made with Intel's 22nm FinFET process [3] - Each P-core is equipped with 3MB of L3 cache, totaling 36MB, and E-core clusters have shared L2 cache, improving inter-core communication [5] - The arrangement of E-cores between P-cores aims to reduce thermal hotspots, with a total of 8 P-cores and 16 E-cores organized strategically [5] Performance Considerations - Despite the innovative chiplet architecture, initial performance has not met expectations, lagging behind AMD's Ryzen 9000 series and even Intel's previous generation processors [6] - Intel is addressing interconnect latency issues through firmware updates, indicating ongoing optimization efforts [6] - The shift to a chiplet architecture is expected to provide future opportunities for architectural enhancements, improving yield and reducing production costs [6]