Workflow
晶圆级芯片
icon
Search documents
清华大学研究团队在晶圆级芯片领域取得重要进展
半导体行业观察· 2025-07-20 04:06
Core Viewpoint - Tsinghua University's research team has made significant advancements in wafer-scale chips, presenting three key research outcomes at the ISCA 2025 conference, focusing on high-performance AI model training and inference scenarios [1][9]. Group 1: Research Achievements - The team developed a collaborative design optimization methodology for wafer-scale chips, integrating computational architecture, integration architecture, and compilation mapping, which has gained recognition in both academia and industry [1][9]. - The research includes a paper on interconnect-centric computational architecture, addressing physical constraints and proposing a "Tick-Tock" co-design framework that optimizes physical and logical topologies [10][12][13]. - Another paper presents a vertically stacked integration architecture that addresses the challenges of tightly coupled heterogeneous design factors, achieving significant improvements in system-level integration density and performance metrics [14][18]. Group 2: Wafer-Scale Chip Technology - Wafer-scale chips represent a disruptive technology that integrates multiple computing, storage, and interconnect components into a single chip, significantly enhancing computational power and efficiency [3][8]. - The design allows for a larger number of transistors to be integrated, overcoming limitations faced by traditional chips, and achieving a chip area of approximately 40,000 square millimeters [4][8]. - The architecture enables higher interconnect density and shorter interconnect distances, resulting in performance and energy efficiency improvements, with potential density reaching over twice that of current supernode solutions [8][9]. Group 3: Industry Context - Major global tech companies, including Tesla and Cerebras Systems, are investing in wafer-scale chip technology, with Tesla's Dojo chip achieving 9 PFlops of computing power and Cerebras' WSE-3 chip integrating 400 trillion transistors [24][25]. - TSMC is also advancing wafer-scale systems, aiming for mass production by 2027, which will enhance computational density and data transfer efficiency [25]. - The advancements in wafer-scale chips are critical for the AI industry's future, as they provide a foundation for high-performance computing necessary for large-scale AI applications [23][26].
晶圆级芯片,是未来
3 6 Ke· 2025-06-29 23:49
Group 1: Industry Overview - The computational power required for large AI models has increased by 1000 times in just two years, significantly outpacing hardware iteration speeds [1] - Current AI training hardware is divided into two main camps: dedicated accelerators using wafer-level integration technology and traditional GPU clusters [1][2] Group 2: Wafer-Level Chips - Wafer-level chips are seen as a breakthrough, allowing for the integration of multiple dies on a single wafer, which enhances bandwidth and reduces latency [3][4] - The size of a single die chip is approximately 858 mm², and the maximum size is constrained by the exposure window [2][3] Group 3: Key Players - Cerebras has developed the WSE-3 wafer-level chip, which utilizes TSMC's 5nm process, featuring 4 trillion transistors and 900,000 AI cores [5][6] - Tesla's Dojo chip employs a different approach, integrating 25 proprietary D1 chips on a wafer, achieving 9 Petaflops of computing power [10][11] Group 4: Performance Comparison - WSE-3 can train models 10 times larger than GPT-4 and Gemini, with a peak performance of 125 PFLOPS [8][14] - In comparison, the WSE-3 has 880 times the on-chip memory capacity and 7000 times the memory bandwidth of the NVIDIA H100 [8][13] Group 5: Cost and Scalability - The cost of Tesla's Dojo system is estimated between $300 million to $500 million, while Cerebras WSE systems range from $2 million to $3 million [18][19] - NVIDIA GPUs, while cheaper initially, face long-term operational cost issues due to high energy consumption and performance bottlenecks [18][19] Group 6: Future Outlook - The wafer-level chip architecture is considered the highest integration density for computing nodes, indicating significant potential for future developments in AI training hardware [20]