存储墙
Search documents
突破“存储墙”,三路并进
3 6 Ke· 2025-12-31 03:35
Core Insights - The explosive growth of AI and high-performance computing is driving an exponential increase in computing demand, leading to a significant challenge known as the "storage wall" [1][2] - The competition for AI and high-performance computing chips will focus not only on transistor density and frequency but also on memory subsystem performance, energy efficiency, and integration innovation [1][4] Group 1: AI and Computing Demand - The evolution of AI models has led to a dramatic increase in computational requirements, with model parameters rising from millions to trillions, resulting in a training computation increase of over 10^18 times in the past 70 years [2][4] - The growth rate of computational performance has significantly outpaced that of memory bandwidth, creating a "bandwidth wall" that limits overall system performance [4][7] Group 2: Memory Technology Challenges - The traditional memory technologies are struggling to meet the unprecedented demands for performance, power consumption, and area (PPA) from various applications, including large language models and edge devices [1][4] - The average growth of DRAM bandwidth over the past 20 years has only been 100 times, compared to a 60,000 times increase in hardware peak floating-point performance [4][7] Group 3: TSMC's Strategic Insights - TSMC emphasizes that the future evolution of memory technology will revolve around "storage-compute synergy," transitioning from traditional on-chip caches to integrated memory solutions that enhance performance and energy efficiency [7][11] - TSMC is focusing on optimizing embedded memory technologies such as SRAM, MRAM, and DCiM to address the challenges posed by AI and HPC demands [11][33] Group 4: SRAM Technology - SRAM is identified as a key technology for high-speed embedded memory, offering low latency, high bandwidth, and low power consumption, making it essential for various high-performance chips [12][16] - The area scaling of SRAM is critical for optimizing chip performance, but it faces challenges as technology nodes advance to 2nm [12][17] Group 5: Computing-in-Memory (CIM) - CIM architecture represents a revolutionary approach that integrates computing capabilities directly into memory arrays, significantly reducing energy consumption and latency associated with data movement [21][24] - TSMC believes that DCiM (Digital Computing-in-Memory) has greater potential than ACiM (Analog Computing-in-Memory) due to its compatibility with advanced processes and flexibility in precision control [26][28] Group 6: MRAM Technology - MRAM is emerging as a viable alternative to traditional embedded flash memory, offering non-volatility, high reliability, and durability, making it suitable for applications in automotive electronics and edge AI [33][35] - TSMC's N16 FinFET embedded MRAM technology meets stringent automotive requirements, showcasing its potential in high-performance applications [39][49] Group 7: System-Level Integration - TSMC advocates for a system-level approach to memory technology breakthroughs, emphasizing the need for 3D packaging and chiplet integration to achieve high bandwidth and low latency [50][54] - The future of AI chips may see a blurring of boundaries between memory and computation, with innovations in 3D stacking and integrated voltage regulators enhancing overall system performance [60][61] Group 8: Future Outlook - The future of storage technology in AI computing is characterized by a comprehensive innovation revolution, with TSMC's roadmap focusing on SRAM, MRAM, and DCiM to overcome the "bandwidth wall" and energy efficiency challenges [62] - The ability to achieve full-stack optimization from transistors to systems will be crucial for leading the next era of AI computing [62]
突破“存储墙”,三路并进
半导体行业观察· 2025-12-31 01:40
Core Viewpoint - The article discusses the exponential growth of AI and high-performance computing, highlighting the emerging challenge of the "storage wall" that limits the performance of AI chips due to inadequate memory bandwidth and efficiency [1][2]. Group 1: AI and Storage Demand - The evolution of AI models has led to a dramatic increase in computational demands, with model parameters rising from millions to trillions, resulting in a training computation increase of over 10^18 times in the past 70 years [2]. - The performance of any computing system is determined by its peak computing power and memory bandwidth, leading to a significant imbalance where hardware peak floating-point performance has increased 60,000 times over the past 20 years, while DRAM bandwidth has only increased 100 times [5][8]. Group 2: Memory Technology Challenges - The rapid growth in computational performance has not been matched by memory bandwidth improvements, creating a "bandwidth wall" that restricts overall system performance [5][8]. - AI inference scenarios are particularly affected, with memory bandwidth becoming a major bottleneck, leading to idle computational resources as they wait for data [8]. Group 3: Future Directions in Memory Technology - TSMC emphasizes that the evolution of memory technology in the AI and HPC era requires a comprehensive optimization across materials, processes, architectures, and packaging [12]. - The future of memory architecture will focus on "storage-compute synergy," transitioning from traditional on-chip caches to integrated memory solutions that enhance performance and efficiency [12][10]. Group 4: SRAM as a Key Technology - SRAM is identified as a critical technology for high-performance embedded memory due to its low latency, high bandwidth, and energy efficiency, widely used in various high-performance chips [13][20]. - TSMC's SRAM technology has evolved through various process nodes, with ongoing innovations aimed at improving density and efficiency [14][22]. Group 5: Computing-in-Memory (CIM) Innovations - CIM architecture represents a revolutionary approach that integrates computing capabilities directly within memory arrays, significantly reducing data movement and energy consumption [23][26]. - TSMC believes that Digital Computing-in-Memory (DCiM) has greater potential than Analog Computing-in-Memory (ACiM) due to its compatibility with advanced processes and flexibility in precision control [28][30]. Group 6: MRAM Developments - MRAM is emerging as a viable alternative to traditional embedded flash memory, offering non-volatility, high reliability, and durability, making it suitable for applications in automotive electronics and edge AI [35][38]. - TSMC's MRAM technology meets stringent automotive requirements, providing robust performance and longevity [41][43]. Group 7: System-Level Integration - TSMC advocates for a system-level approach to memory and compute integration, utilizing advanced packaging technologies like 2.5D/3D integration to enhance bandwidth and reduce latency [50][52]. - The future of AI chips may see a blurring of the lines between memory and compute, with tightly integrated architectures that optimize energy efficiency and performance [58][60].
智能早报丨“大空头”做空英伟达与Palantir;苹果中国严控线下经销商线上销售
Guan Cha Zhe Wang· 2025-11-05 02:16
Group 1: Investment Actions - Michael Burry's Scion Asset Management has significantly shorted NVIDIA and Palantir, with these positions making up 80% of its investment portfolio [1][3] - The nominal value of put options for Palantir is $912.1 million (5 million shares), while for NVIDIA it is $186.6 million (1 million shares) [4] - Burry's actions align with his previous warnings about an AI bubble, drawing parallels to the 1999-2000 internet bubble [5] Group 2: Company Responses - Palantir's CEO Alex Karp criticized Burry's shorting strategy, arguing that both Palantir and NVIDIA are highly profitable companies [3] - Burry's short positions are currently facing losses as both NVIDIA and Palantir's stock prices have risen since the end of September [5] Group 3: Market Trends and Insights - The AI storage market is experiencing structural changes due to demand driven by AI computing needs, with SK Hynix announcing new AI storage products [14][15] - The robotics industry in China has seen a revenue increase of 29.5% in the first three quarters of the year, driven by manufacturing upgrades and new industry demands [16]
一文看懂“存算一体”
Hu Xiu· 2025-08-15 06:52
Core Concept - The article discusses the concept of "Compute In Memory" (CIM), which integrates storage and computation to enhance data processing efficiency and reduce energy consumption [1][20]. Group 1: Background and Need for CIM - Traditional computing architecture, known as the von Neumann architecture, separates storage and computation, leading to inefficiencies as data transfer speeds cannot keep up with processing speeds [2][10]. - The explosion of data in the internet era and the rise of AI have highlighted the limitations of this architecture, resulting in the emergence of the "memory wall" and "power wall" challenges [11][12]. - The "memory wall" refers to the inadequate data transfer speeds between storage and processors, while the "power wall" indicates high energy consumption during data transfer [13][16]. Group 2: Development of CIM - Research on CIM dates back to 1969, but significant advancements have only occurred in the 21st century due to improvements in chip and semiconductor technologies [23][26]. - Notable developments include the use of memristors for logic functions and the construction of CIM architectures for deep learning, which can achieve significant reductions in power consumption and increases in speed [27][28]. - The recent surge in AI demands has accelerated the development of CIM technologies, with numerous startups entering the field alongside established chip manufacturers [30][31]. Group 3: Technical Classification of CIM - CIM is categorized into three types based on the proximity of storage and computation: Processing Near Memory (PNM), Processing In Memory (PIM), and Computing In Memory (CIM) [34][35]. - PNM involves integrating storage and computation units to enhance data transfer efficiency, while PIM integrates computation capabilities directly into memory chips [36][40]. - CIM represents the true integration of storage and computation, eliminating the distinction between the two and allowing for efficient data processing directly within storage units [43][46]. Group 4: Applications of CIM - CIM is particularly suited for AI-related computations, including natural language processing and intelligent decision-making, where efficiency and energy consumption are critical [61][62]. - It also has potential applications in AIoT products and high-performance cloud computing scenarios, where traditional architectures struggle to meet diverse computational needs [63][66]. Group 5: Market Potential and Challenges - The global CIM technology market is projected to reach $30.63 billion by 2029, with a compound annual growth rate (CAGR) of 154.7% [79]. - Despite its potential, CIM faces technical challenges related to semiconductor processes and the establishment of a supportive ecosystem for design and testing tools [70][72]. - Market challenges include competition with traditional architectures and the need for cost-effective solutions that meet user demands [74][76].
DRAM“危机”
半导体行业观察· 2025-04-20 03:50
Core Viewpoint - The article discusses the rapid advancements in AI and the challenges posed by the "memory wall" problem, highlighting the need for innovative storage solutions to meet the increasing demands of AI models and high-performance computing [1][2]. Group 1: Memory Wall and HBM Technology - The growth of AI models has led to an exponential increase in model parameters, creating significant demands on computing resources, particularly storage bandwidth [1]. - Traditional DRAM bandwidth growth is lagging behind processor performance, with DRAM bandwidth increasing only 1.6 times every two years compared to processor performance increasing threefold [1]. - HBM technology has emerged as a revolutionary solution, offering data transfer speeds of 1.2TB per second, significantly alleviating memory bandwidth pressure [2]. Group 2: 3D Ferroelectric RAM - 3D Ferroelectric RAM (FeRAM) is highlighted as a potential disruptor in the DRAM landscape, with companies like SunRise Memory developing innovative FeFET storage units that promise tenfold storage density improvements over traditional DRAM [4][5]. - This new technology boasts a 90% reduction in power consumption compared to traditional DRAM, making it particularly advantageous for energy-sensitive AI applications [5]. - SunRise Memory aims to leverage existing 3D NAND fabrication processes for mass production, indicating a strategic approach to commercialization [5][6]. Group 3: Other Emerging Storage Technologies - Neumonda GmbH and Ferroelectric Memory Co. are collaborating to develop "DRAM+" non-volatile memory, which integrates ferroelectric effects to create low-power, high-performance storage solutions [8][9]. - Imec's 2T0C DRAM architecture represents a significant innovation, allowing for higher density and improved performance by eliminating the need for capacitors [10][11]. - Phase Change Memory (PCM) is also gaining traction, with advancements in nanowire technology reducing power consumption significantly while maintaining high performance [19][20]. Group 4: Market Outlook and Industry Implications - The semiconductor industry is undergoing a transformation driven by AI, with various new storage technologies vying to replace traditional DRAM [25]. - The emergence of diverse storage solutions, including 3D Ferroelectric RAM, DRAM+, and IGZO 2T0C, indicates a shift towards a more versatile storage market that can cater to different application needs [25]. - The ongoing developments in storage technology are expected to reshape the semiconductor landscape, presenting both opportunities and challenges for industry players [25].