存储墙
Search documents
突破“存储墙”,三路并进
3 6 Ke· 2025-12-31 03:35
前言 近年来,AI与高性能计算的爆发式增长,正推动计算需求呈指数级攀升。从ChatGPT的横空出世到Sora带来的视觉震撼,大规模AI模型不仅在参数规模上 指数级膨胀,其对计算能力的需求更是呈现出令人惊叹的增长曲线。 然而,在这片繁荣的背后,一个日益严峻的挑战正浮出水面——"存储墙"。 从千亿参数的大语言模型到边缘端的智能终端,各类应用对存储器的性能、功耗、面积(PPA)提出了前所未有的严苛要求。存储"带宽墙"成为制约AI计 算吞吐量与延迟的核心瓶颈,传统存储器技术已难以满足系统能效优化需求,巨大的性能缺口正制约着AI芯片发挥其全部潜力。 作为全球半导体制造的领导者,台积电深刻洞察到这一根本性矛盾。在2025年的IEDM(国际电子器件会议)教程中,台积电清晰指出:未来AI与高性能 计算芯片的竞争,将不仅仅是晶体管密度与频率的竞赛,更是内存子系统性能、能效与集成创新的综合较量。 AI算力狂奔下,存储"带宽墙"成核心痛点 AI模型的进化史,堪称一场对算力与存储的极限压榨。 从早期的AlexNet到如今的GPT-4、Llama2、PaLM,模型参数从百万级跃升至万亿级,模型规模的扩张直接带动训练与推理阶段的计算量( ...
突破“存储墙”,三路并进
半导体行业观察· 2025-12-31 01:40
Core Viewpoint - The article discusses the exponential growth of AI and high-performance computing, highlighting the emerging challenge of the "storage wall" that limits the performance of AI chips due to inadequate memory bandwidth and efficiency [1][2]. Group 1: AI and Storage Demand - The evolution of AI models has led to a dramatic increase in computational demands, with model parameters rising from millions to trillions, resulting in a training computation increase of over 10^18 times in the past 70 years [2]. - The performance of any computing system is determined by its peak computing power and memory bandwidth, leading to a significant imbalance where hardware peak floating-point performance has increased 60,000 times over the past 20 years, while DRAM bandwidth has only increased 100 times [5][8]. Group 2: Memory Technology Challenges - The rapid growth in computational performance has not been matched by memory bandwidth improvements, creating a "bandwidth wall" that restricts overall system performance [5][8]. - AI inference scenarios are particularly affected, with memory bandwidth becoming a major bottleneck, leading to idle computational resources as they wait for data [8]. Group 3: Future Directions in Memory Technology - TSMC emphasizes that the evolution of memory technology in the AI and HPC era requires a comprehensive optimization across materials, processes, architectures, and packaging [12]. - The future of memory architecture will focus on "storage-compute synergy," transitioning from traditional on-chip caches to integrated memory solutions that enhance performance and efficiency [12][10]. Group 4: SRAM as a Key Technology - SRAM is identified as a critical technology for high-performance embedded memory due to its low latency, high bandwidth, and energy efficiency, widely used in various high-performance chips [13][20]. - TSMC's SRAM technology has evolved through various process nodes, with ongoing innovations aimed at improving density and efficiency [14][22]. Group 5: Computing-in-Memory (CIM) Innovations - CIM architecture represents a revolutionary approach that integrates computing capabilities directly within memory arrays, significantly reducing data movement and energy consumption [23][26]. - TSMC believes that Digital Computing-in-Memory (DCiM) has greater potential than Analog Computing-in-Memory (ACiM) due to its compatibility with advanced processes and flexibility in precision control [28][30]. Group 6: MRAM Developments - MRAM is emerging as a viable alternative to traditional embedded flash memory, offering non-volatility, high reliability, and durability, making it suitable for applications in automotive electronics and edge AI [35][38]. - TSMC's MRAM technology meets stringent automotive requirements, providing robust performance and longevity [41][43]. Group 7: System-Level Integration - TSMC advocates for a system-level approach to memory and compute integration, utilizing advanced packaging technologies like 2.5D/3D integration to enhance bandwidth and reduce latency [50][52]. - The future of AI chips may see a blurring of the lines between memory and compute, with tightly integrated architectures that optimize energy efficiency and performance [58][60].
智能早报丨“大空头”做空英伟达与Palantir;苹果中国严控线下经销商线上销售
Guan Cha Zhe Wang· 2025-11-05 02:16
Group 1: Investment Actions - Michael Burry's Scion Asset Management has significantly shorted NVIDIA and Palantir, with these positions making up 80% of its investment portfolio [1][3] - The nominal value of put options for Palantir is $912.1 million (5 million shares), while for NVIDIA it is $186.6 million (1 million shares) [4] - Burry's actions align with his previous warnings about an AI bubble, drawing parallels to the 1999-2000 internet bubble [5] Group 2: Company Responses - Palantir's CEO Alex Karp criticized Burry's shorting strategy, arguing that both Palantir and NVIDIA are highly profitable companies [3] - Burry's short positions are currently facing losses as both NVIDIA and Palantir's stock prices have risen since the end of September [5] Group 3: Market Trends and Insights - The AI storage market is experiencing structural changes due to demand driven by AI computing needs, with SK Hynix announcing new AI storage products [14][15] - The robotics industry in China has seen a revenue increase of 29.5% in the first three quarters of the year, driven by manufacturing upgrades and new industry demands [16]
一文看懂“存算一体”
Hu Xiu· 2025-08-15 06:52
Core Concept - The article discusses the concept of "Compute In Memory" (CIM), which integrates storage and computation to enhance data processing efficiency and reduce energy consumption [1][20]. Group 1: Background and Need for CIM - Traditional computing architecture, known as the von Neumann architecture, separates storage and computation, leading to inefficiencies as data transfer speeds cannot keep up with processing speeds [2][10]. - The explosion of data in the internet era and the rise of AI have highlighted the limitations of this architecture, resulting in the emergence of the "memory wall" and "power wall" challenges [11][12]. - The "memory wall" refers to the inadequate data transfer speeds between storage and processors, while the "power wall" indicates high energy consumption during data transfer [13][16]. Group 2: Development of CIM - Research on CIM dates back to 1969, but significant advancements have only occurred in the 21st century due to improvements in chip and semiconductor technologies [23][26]. - Notable developments include the use of memristors for logic functions and the construction of CIM architectures for deep learning, which can achieve significant reductions in power consumption and increases in speed [27][28]. - The recent surge in AI demands has accelerated the development of CIM technologies, with numerous startups entering the field alongside established chip manufacturers [30][31]. Group 3: Technical Classification of CIM - CIM is categorized into three types based on the proximity of storage and computation: Processing Near Memory (PNM), Processing In Memory (PIM), and Computing In Memory (CIM) [34][35]. - PNM involves integrating storage and computation units to enhance data transfer efficiency, while PIM integrates computation capabilities directly into memory chips [36][40]. - CIM represents the true integration of storage and computation, eliminating the distinction between the two and allowing for efficient data processing directly within storage units [43][46]. Group 4: Applications of CIM - CIM is particularly suited for AI-related computations, including natural language processing and intelligent decision-making, where efficiency and energy consumption are critical [61][62]. - It also has potential applications in AIoT products and high-performance cloud computing scenarios, where traditional architectures struggle to meet diverse computational needs [63][66]. Group 5: Market Potential and Challenges - The global CIM technology market is projected to reach $30.63 billion by 2029, with a compound annual growth rate (CAGR) of 154.7% [79]. - Despite its potential, CIM faces technical challenges related to semiconductor processes and the establishment of a supportive ecosystem for design and testing tools [70][72]. - Market challenges include competition with traditional architectures and the need for cost-effective solutions that meet user demands [74][76].
DRAM“危机”
半导体行业观察· 2025-04-20 03:50
Core Viewpoint - The article discusses the rapid advancements in AI and the challenges posed by the "memory wall" problem, highlighting the need for innovative storage solutions to meet the increasing demands of AI models and high-performance computing [1][2]. Group 1: Memory Wall and HBM Technology - The growth of AI models has led to an exponential increase in model parameters, creating significant demands on computing resources, particularly storage bandwidth [1]. - Traditional DRAM bandwidth growth is lagging behind processor performance, with DRAM bandwidth increasing only 1.6 times every two years compared to processor performance increasing threefold [1]. - HBM technology has emerged as a revolutionary solution, offering data transfer speeds of 1.2TB per second, significantly alleviating memory bandwidth pressure [2]. Group 2: 3D Ferroelectric RAM - 3D Ferroelectric RAM (FeRAM) is highlighted as a potential disruptor in the DRAM landscape, with companies like SunRise Memory developing innovative FeFET storage units that promise tenfold storage density improvements over traditional DRAM [4][5]. - This new technology boasts a 90% reduction in power consumption compared to traditional DRAM, making it particularly advantageous for energy-sensitive AI applications [5]. - SunRise Memory aims to leverage existing 3D NAND fabrication processes for mass production, indicating a strategic approach to commercialization [5][6]. Group 3: Other Emerging Storage Technologies - Neumonda GmbH and Ferroelectric Memory Co. are collaborating to develop "DRAM+" non-volatile memory, which integrates ferroelectric effects to create low-power, high-performance storage solutions [8][9]. - Imec's 2T0C DRAM architecture represents a significant innovation, allowing for higher density and improved performance by eliminating the need for capacitors [10][11]. - Phase Change Memory (PCM) is also gaining traction, with advancements in nanowire technology reducing power consumption significantly while maintaining high performance [19][20]. Group 4: Market Outlook and Industry Implications - The semiconductor industry is undergoing a transformation driven by AI, with various new storage technologies vying to replace traditional DRAM [25]. - The emergence of diverse storage solutions, including 3D Ferroelectric RAM, DRAM+, and IGZO 2T0C, indicates a shift towards a more versatile storage market that can cater to different application needs [25]. - The ongoing developments in storage technology are expected to reshape the semiconductor landscape, presenting both opportunities and challenges for industry players [25].