Workflow
AI内存压缩技术
icon
Search documents
存储价格崩盘?记者实探华强北
新华网财经· 2026-04-01 06:45
Core Viewpoint - The recent decline in memory prices is attributed to market fluctuations, but the overall supply shortage remains unchanged, with major clients accelerating purchases to mitigate ongoing shortages [1][4][9]. Group 1: Market Dynamics - The memory market has experienced a price drop of 300 to 500 yuan compared to the peak at the end of January, but the notion of widespread panic selling among individual merchants is not entirely accurate [1][4]. - The price of DDR5 32G memory modules is influenced by various factors, with current market prices for popular models ranging from 2600 to 2800 yuan, reflecting a decrease of 300 to 500 yuan from previous highs [6][8]. - Reports of panic selling in the market are linked to a new AI memory compression technology introduced by Google, which has caused some fear among traders, although the actual impact on prices has been limited [5][9]. Group 2: Supply and Demand - The storage market is expected to remain in a state of "tight balance" or even "hard shortage" for at least the next 24 months, driven by the increasing demand for AI applications [11]. - The core driver of the current price surge is the shift in demand from traditional consumer electronics to AI-related needs, which has created a structural mismatch between supply and demand [9][10]. - Analysts predict that DRAM prices will continue to rise until 2027, primarily due to high visibility in data center infrastructure demand from cloud service providers [10].
谷歌发布KV缓存压缩技术,存储需求预期遭冲击,美股存储板块集体下挫!
美股IPO· 2026-03-25 23:04
Core Viewpoint - Google's TurboQuant memory compression technology significantly reduces the memory footprint of large language models, raising concerns about storage demand in the market [1][3][5]. Group 1: Technology Overview - TurboQuant can compress key-value cache to 3 bits, achieving a 6-fold reduction in memory usage and up to 8 times acceleration without requiring model retraining or fine-tuning [6]. - The technology utilizes a two-step process: first, it applies the PolarQuant method for high-quality compression, followed by the Johnson-Lindenstrauss algorithm to eliminate residual errors [6]. - TurboQuant has been validated in various benchmark tests and is expected to be presented at ICLR 2026, while PolarQuant is planned for AISTATS 2026 [6]. Group 2: Market Reaction - Following the announcement of TurboQuant, the storage chip sector experienced significant declines, with companies like SanDisk and Micron dropping over 3.4%, and Seagate and Western Digital also facing losses [2][3]. - The storage chip and hardware supply chain index fell by 2.08%, reflecting market concerns about future storage demand [3]. Group 3: Implications for AI Applications - Morgan Stanley notes that TurboQuant primarily affects the inference stage and does not reduce the high bandwidth memory (HBM) used for model weights, suggesting that overall storage demand may not decrease by 6 times [7]. - The efficiency gains from TurboQuant could lead to increased throughput on existing hardware, allowing for longer context lengths and larger batch sizes without triggering memory overflow [7]. - The technology may lower the service costs of AI deployments, making it more profitable and potentially enabling more applications to emerge, thus increasing infrastructure utilization [7][8]. Group 4: Long-term Outlook - Morgan Stanley describes TurboQuant as a breakthrough that could reshape the cost curve of AI deployment, providing positive signals for cloud service providers and model platforms [8]. - The long-term impact on computing and memory hardware is assessed as neutral to slightly positive, indicating a balanced view on future investments in the sector [8].