Workflow
AI存储,再度爆火

Core Viewpoint - The rapid development of AI has made storage a critical component in the AI infrastructure, alongside computing power. The demand for storage is surging due to the increasing data volume and inference scenarios driven by large models and generative AI. Three storage technologies—HBM, HBF, and GDDR7—are redefining the future landscape of AI infrastructure [1]. Group 1: HBM (High Bandwidth Memory) - HBM has evolved from a high-performance AI chip component to a strategic point in the storage industry, significantly impacting AI chip performance limits. In less than three years, HBM has achieved over twofold capacity and approximately 2.5 times bandwidth increase [3]. - SK Hynix is leading the HBM market, currently in the final testing phase for the sixth generation (HBM4) and has announced readiness for mass production. In contrast, Samsung is facing challenges in HBM4 supply to Nvidia, with a two-month delay in testing [3][5]. - A notable trend is the customization of HBM, driven by cloud giants developing their AI chips. SK Hynix is shifting towards a fully customized HBM approach, collaborating closely with major clients [4]. Group 2: HBF (High Bandwidth Flash) - HBF aims to address the limitations of traditional storage by combining the capacity of NAND flash with the bandwidth of HBM. Sandisk is leading the development of HBF technology, which is expected to meet the growing storage demands of AI applications [8][9]. - HBF is seen as complementary to HBM, suitable for specific applications requiring large block storage units. It is particularly advantageous in scenarios demanding high capacity but with relatively relaxed bandwidth requirements [10][11]. Group 3: GDDR7 - Nvidia's introduction of the Rubin CPX GPU, utilizing GDDR7 instead of HBM4, reflects a new approach to AI inference architecture. This design optimizes resource allocation by separating the inference process into two stages, effectively utilizing GDDR7 for context building [13]. - The demand for GDDR7 is increasing, with Samsung successfully meeting Nvidia's orders. This flexibility positions Samsung favorably in the graphics DRAM market [14]. - GDDR7's cost-effectiveness may drive the widespread adoption of AI inference infrastructure, potentially increasing overall market demand for high-end HBM due to the proliferation of applications [15]. Group 4: Industry Trends and Future Outlook - The collaborative evolution of storage technologies is crucial for the AI industry's growth. HBM remains essential for high-end training and inference, while HBF and GDDR7 cater to diverse market needs [23]. - The ongoing innovation in storage technology will accelerate as AI applications expand across various sectors, providing tailored solutions for both performance-driven and cost-sensitive users [23].