Workflow
存算一体
icon
Search documents
HBM,挑战加倍
3 6 Ke· 2025-08-19 10:59
Core Insights - High Bandwidth Memory (HBM) is emerging as a critical component for AI model training and inference due to its unique 3D stacked structure, which significantly enhances data transfer rates compared to traditional memory solutions like GDDR [1][2] HBM Market Dynamics - SK Hynix has established a dominant position in the HBM market, with its market share surpassing that of Samsung, which has seen a decline from 41% to 17% in the same period [3][4] - The launch of HBM3E has been pivotal for SK Hynix, attracting major tech companies like AMD, NVIDIA, Microsoft, and Amazon, leading to a significant increase in demand [3] - SK Hynix's sales in DRAM and NAND reached approximately 21.8 trillion KRW, surpassing Samsung's 21.2 trillion KRW for the first time [3] Competitive Landscape - Samsung is attempting to regain its footing by reviving Z-NAND technology, aiming for performance improvements of up to 15 times over traditional NAND and a reduction in power consumption by up to 80% [6][7] - NEO Semiconductor has introduced X-HBM architecture, which offers 16 times the bandwidth and 10 times the density of existing memory technologies, targeting the AI chip market [10] - Saimemory, a collaboration between SoftBank, Intel, and Tokyo University, is developing a new stacked DRAM architecture aimed at becoming a direct HBM alternative with significant performance improvements [11] Innovations and Alternatives - SanDisk and SK Hynix are collaborating on High Bandwidth Flash (HBF), a new storage architecture designed for AI applications, which combines 3D NAND flash with HBM characteristics [12][13] - The industry is exploring various architectural innovations, such as Processing-In-Memory (PIM), to reduce reliance on HBM and enhance efficiency [15][16] Future Trends - The AI memory market is expected to evolve into a heterogeneous multi-tiered structure, where HBM will focus on training scenarios, while PIM memory will cater to high-efficiency inference applications [18] - The demand for HBM, particularly HBM3 and above, is projected to remain strong, with significant price increases noted in the market [17]
一文看懂“存算一体”
Hu Xiu· 2025-08-15 06:52
Core Concept - The article discusses the concept of "Compute In Memory" (CIM), which integrates storage and computation to enhance data processing efficiency and reduce energy consumption [1][20]. Group 1: Background and Need for CIM - Traditional computing architecture, known as the von Neumann architecture, separates storage and computation, leading to inefficiencies as data transfer speeds cannot keep up with processing speeds [2][10]. - The explosion of data in the internet era and the rise of AI have highlighted the limitations of this architecture, resulting in the emergence of the "memory wall" and "power wall" challenges [11][12]. - The "memory wall" refers to the inadequate data transfer speeds between storage and processors, while the "power wall" indicates high energy consumption during data transfer [13][16]. Group 2: Development of CIM - Research on CIM dates back to 1969, but significant advancements have only occurred in the 21st century due to improvements in chip and semiconductor technologies [23][26]. - Notable developments include the use of memristors for logic functions and the construction of CIM architectures for deep learning, which can achieve significant reductions in power consumption and increases in speed [27][28]. - The recent surge in AI demands has accelerated the development of CIM technologies, with numerous startups entering the field alongside established chip manufacturers [30][31]. Group 3: Technical Classification of CIM - CIM is categorized into three types based on the proximity of storage and computation: Processing Near Memory (PNM), Processing In Memory (PIM), and Computing In Memory (CIM) [34][35]. - PNM involves integrating storage and computation units to enhance data transfer efficiency, while PIM integrates computation capabilities directly into memory chips [36][40]. - CIM represents the true integration of storage and computation, eliminating the distinction between the two and allowing for efficient data processing directly within storage units [43][46]. Group 4: Applications of CIM - CIM is particularly suited for AI-related computations, including natural language processing and intelligent decision-making, where efficiency and energy consumption are critical [61][62]. - It also has potential applications in AIoT products and high-performance cloud computing scenarios, where traditional architectures struggle to meet diverse computational needs [63][66]. Group 5: Market Potential and Challenges - The global CIM technology market is projected to reach $30.63 billion by 2029, with a compound annual growth rate (CAGR) of 154.7% [79]. - Despite its potential, CIM faces technical challenges related to semiconductor processes and the establishment of a supportive ecosystem for design and testing tools [70][72]. - Market challenges include competition with traditional architectures and the need for cost-effective solutions that meet user demands [74][76].
对话「后摩智能」吴强:从科学家到创业者的惊险一跃
3 6 Ke· 2025-08-06 00:02
Core Insights - The article highlights the significant advancements in China's computing power sector, particularly focusing on "super nodes" and edge AI chips as key trends in the AI landscape [1][2] - The emergence of edge computing is seen as a potential larger market than cloud computing, with companies like Houmo Intelligence positioned to capitalize on this opportunity [2][3] - Houmo Intelligence's M50 chip, based on in-memory computing technology, represents a breakthrough in efficiency and performance for edge AI applications [3][6] Group 1: Industry Trends - The development of large AI models has created a strong demand for cloud computing, while edge computing is gaining traction due to its ability to reduce computational needs for generative AI applications [1][2] - The CEO of Houmo Intelligence predicts that 90% of data processing for generative AI will occur at the edge, with only 10% requiring cloud resources [1][2] - The market for edge computing is expected to accommodate more players, potentially leading to the emergence of the "next Nvidia" [2] Group 2: Company Overview - Houmo Intelligence, founded by CEO Wu Qiang, focuses on in-memory computing technology to enhance AI chip efficiency, having transitioned from an initial focus on smart driving chips to general-purpose edge AI applications [2][8] - The M50 chip features significant performance metrics, including 160 TOPS@INT8 and 100 TFLOPS@bFP16, with a typical power consumption of only 10W, making it suitable for various smart devices [6][7] - The company has established partnerships with notable clients, including Lenovo and iFlytek, to expand its market presence in edge AI applications [7][10] Group 3: Technological Innovations - The M50 chip utilizes a new architecture called "Tianxuan" IPU, which allows floating-point models to run directly on the in-memory computing architecture, enhancing application efficiency [6][7] - The in-memory computing approach addresses the "memory wall" and "power wall" issues associated with traditional computing architectures, making it a promising solution for future AI applications [2][3] - The company has developed a new compiler toolchain, "Houmo Dadao," to facilitate easy adaptation of its chips to mainstream deep learning frameworks [6][15] Group 4: Market Dynamics - The edge AI chip market is characterized by cost sensitivity, power efficiency, and compact design requirements, which are critical for successful product deployment [11][12] - The transition from cloud to edge computing is driven by the need for high efficiency and low power consumption in AI applications, particularly in consumer electronics and smart devices [10][11] - The competitive landscape is evolving, with various companies exploring in-memory computing, leading to a diverse range of approaches and technologies in the market [12][13]
对话「后摩智能」吴强:从科学家到创业者的惊险一跃
36氪· 2025-08-05 13:49
Core Viewpoint - The article emphasizes the significance of "storage-compute integration" as a key technology for edge AI chips, which is expected to revolutionize the last mile of large model computing, enabling efficient local processing and reducing reliance on cloud computing [2][4][6]. Group 1: Industry Trends - The AI model development has led to a two-tiered growth in computing power, with cloud computing expanding for model training and edge AI chips gaining traction for inference applications [4][5]. - The emergence of "super nodes" and edge AI chips was highlighted at WAIC 2025, showcasing the growing importance of localized computing solutions [3][4]. - The market for edge computing is anticipated to be larger than cloud computing, presenting opportunities for new players to emerge, potentially creating the "next Nvidia" [4][5]. Group 2: Company Insights - The company, Houmo Intelligent, founded by CEO Wu Qiang, focuses on developing AI chips based on storage-compute integration technology, aiming to address the challenges of traditional computing architectures [5][6]. - The newly launched M50 chip utilizes innovative architecture and compiler tools to enhance efficiency and ease of use, supporting mainstream deep learning frameworks [8][10]. - The M50 chip boasts impressive specifications, achieving 160 TOPS@INT8 and 100 TFLOPS@bFP16 with a power consumption of only 10W, making it suitable for various smart devices without cloud dependency [8][10]. Group 3: Market Strategy - The company is targeting multiple application areas, including consumer electronics, smart voice systems, and edge computing for telecom operators, with notable interest from clients like Lenovo and China Mobile [14][15]. - The transition from a focus on smart driving chips to general-purpose edge AI chips reflects a strategic pivot in response to market demands and opportunities in large model applications [11][13]. - The company aims to leverage its expertise in storage-compute integration to meet the growing needs for efficient AI processing in diverse sectors [17][18].
商道创投网·会员动态|燕芯微电子·完成近亿元天使轮融资
Sou Hu Cai Jing· 2025-08-04 13:19
《商道创投网》创业家会员·本轮融资用途是什么? 燕芯微董事长兼首席科学家表示:本轮资金将主要用于三大方向——一是持续加码ReRAM器件与工艺 研发,加速高密度阵列良率提升;二是扩充存算一体AI芯片研发团队,推动首批客户验证;三是搭建 开放生态实验室,与上下游伙伴共建国产新型存储标准。 《商道创投网》创投家会员·本轮投资原因是什么? 领航新界创始合伙人表示:我们看中燕芯微在ReRAM赛道十年磨一剑的原创技术壁垒,其器件一致 性、阵列集成度已达到国际第一梯队水平;同时北大系团队在工艺、设计、商业化上的复合能力,使公 司具备快速迭代与规模落地的双重潜力,有望填补国内新型存储产业化空白。 《商道创投网》2025年8月4日从官方获悉:燕芯微电子(上海)有限公司近日完成了由领航新界、燕缘 创投联合领投,考拉基金、芯阳基金、华宇科创投、佰维存储等机构共同参与的近亿元天使轮融资。 《商道创投网》创投生态圈·本轮投融观点是什么? 《商道创投网》创业家会员·单位简介: 商道创投网创始人王帥表示:本轮融资恰逢国家"创投十七条"与上海浦东引领区新政叠加落地,政府资 金、产业资本、高校成果三方共振;燕芯微作为北大科研成果转化标杆,既验 ...
AI算力集群迈进“万卡”时代 超节点为什么火了?
Di Yi Cai Jing· 2025-07-30 10:24
Core Insights - The recent WAIC showcased the rising trend of supernodes, with multiple companies, including Huawei and Shanghai Yidian, presenting their supernode solutions, indicating a growing interest in high-performance computing [1][2][4] Group 1: Supernode Technology - Supernodes are designed to address the challenges of large-scale computing clusters by integrating computing resources to enhance efficiency and support models with trillions of parameters [1][2] - The technology allows for improved performance even when individual chip manufacturing processes are limited, marking a significant trend in the industry [1][5] - Supernodes can be developed through two main approaches: scale-out (horizontal expansion) and scale-up (vertical expansion), optimizing communication bandwidth and latency within the nodes [3][4] Group 2: Market Dynamics - The share of domestic AI chips in AI servers is increasing, with projections indicating a drop in reliance on foreign chips from 63% to 49% this year [6] - Companies like Nvidia are still focusing on the Chinese market, indicating the competitive landscape remains intense [6] - Domestic manufacturers are exploring alternative strategies to compete with established players like Nvidia, including optimizing for specific applications such as AI inference [6][8] Group 3: Innovation in Chip Design - Some domestic chip manufacturers are adopting sparse computing techniques, which require less stringent manufacturing processes, allowing for broader applicability in various scenarios [7] - Companies are focusing on edge computing and AI inference, aiming to reduce costs and improve efficiency in specific applications [8] - The introduction of new chips, such as the Homa M50, highlights the industry's shift towards innovative solutions that leverage emerging technologies like in-memory computing [8]
最高能效比!他又死磕“存算一体”2年,拿出全新端边大模型AI芯片
量子位· 2025-07-28 06:42
Core Viewpoint - The article highlights the launch of the M50 AI chip by Houmo Intelligent, which boasts the highest energy efficiency in the industry for integrated storage and computing, marking a significant advancement in AI technology [3][4][8]. Group 1: Product Launch and Specifications - The M50 chip features 160 TOPS@INT8 physical computing power, 100 TFLOPS@bFP16 floating-point computing power, and a bandwidth of 153.6 GB/s, with a typical power consumption of only 10W [4][8]. - The M50 is built on the second-generation integrated storage and computing technology developed by Houmo Intelligent, which allows for significant improvements in energy efficiency [8][9]. Group 2: Technological Innovation - The integrated storage and computing technology merges computation and storage, eliminating the need for data transfer between memory and processing units, thus overcoming the "power wall" and "storage wall" limitations of traditional architectures [11][12]. - The M50 utilizes SRAM-CIM technology, which involves deep structural changes to SRAM arrays, enabling parallel loading and computation, thereby doubling efficiency [12][15]. Group 3: Software and Ecosystem - Accompanying the M50 is the new compiler toolchain, Houmo Avenue®, which simplifies the optimization process for developers, allowing for automatic search of the best strategies [24]. - The company has developed a complete product matrix that includes various hardware solutions for both terminal and edge computing, enhancing the accessibility of AI capabilities across different applications [28][36]. Group 4: Market Positioning and Future Outlook - Houmo Intelligent's focus on integrated storage and computing is seen as a necessary differentiation strategy in a competitive landscape dominated by giants like NVIDIA and Huawei [37][40]. - The company aims to address the increasing demand for computing power and bandwidth in the era of large models, with a vision of making AI capabilities ubiquitous in everyday devices [41][42].
Jinqiu Spotlight | 锦秋基金被投光本位研发全球首颗存算一体光芯片
锦秋集· 2025-07-22 15:04
Core Viewpoint - The article discusses the strategic investment by Jinqiu Capital in "Guangbenwei Technology," a company specializing in optical computing chips, highlighting its innovative technology and market potential in the AI sector [2][20]. Company Overview - Guangbenwei Technology was founded by two young entrepreneurs who returned to China to establish the company after gaining experience abroad. The company has developed the world's first optical computing chip that meets commercial standards for computing density and precision [4][7]. - The founders, Xiong Yingjiang and Cheng Tangsheng, have extensive backgrounds in AI and optical computing, which they leveraged to create a unique product that integrates optical technology with computing capabilities [4][6]. Technology and Innovation - Guangbenwei Technology has achieved significant milestones, including the successful development of a 128x128 matrix optical computing chip, which is the first of its kind to integrate storage and computing functions [10][12]. - The company utilizes a unique technology route that combines silicon photonics with phase change materials (PCM), allowing for a significant reduction in energy consumption and an increase in computing power [13][14]. - The optical chips developed by Guangbenwei can potentially offer over 1000 times the computing power of traditional electronic chips while consuming less energy, addressing the growing demand for computational power in AI applications [8][14]. Market Demand and Applications - The demand for computing power is expected to surge, with global data centers projected to consume approximately 415 terawatt-hours of electricity in 2024, potentially doubling by 2030 [7]. - Guangbenwei Technology targets two main customer segments: large internet companies with advanced computing capabilities and government-led intelligent computing centers, each with distinct needs for energy efficiency and economic viability [16][17]. Funding and Growth - Guangbenwei Technology has successfully completed multiple funding rounds, including a strategic round led by Jinqiu Capital, which reflects investor confidence in the company's technology and market potential [2][20]. - The company is actively collaborating with leading internet firms, GPU manufacturers, and research institutions to validate its technology and expand its market presence [19].
两位95后创立光计算芯片公司,研发全球首颗存算一体光芯片
3 6 Ke· 2025-07-22 02:28
Core Insights - The article discusses the establishment and progress of a company called "Guangbenwei," which has developed the world's first optical computing chip that meets commercial standards in terms of computing density and precision [1][2] - The founders, Xiong Yingjiang and Cheng Tangsheng, have extensive backgrounds in AI and optical computing, and they identified a significant opportunity in the optical computing sector due to the increasing demand for computing power driven by AI advancements [2][3] Company Overview - Guangbenwei was founded by two young entrepreneurs who previously had experience in the U.S. and academia, focusing on the development of optical computing chips that integrate silicon photonics and phase change materials [1][3] - The company has achieved significant milestones, including the successful tape-out of a 128x128 matrix optical computing chip, making it the only company to achieve such integration on a single die [1][7] Market Context - The demand for computing power is expected to surge, with global data centers projected to consume approximately 415 terawatt-hours of electricity in 2024, potentially doubling by 2030 [2] - Optical chips are believed to have the potential to outperform traditional electronic chips by over 1000 times in terms of computing power while consuming less energy [3] Technological Advancements - Guangbenwei's optical computing chips utilize a unique architecture that allows for a programmable structure with over 16,000 adjustable nodes, making it adaptable to various model parameters [7][9] - The integration of storage and computing functions within the chip significantly alleviates storage pressure and enhances performance [9] Commercial Strategy - The company targets two main customer segments: large internet companies with advanced computing capabilities and government-led intelligent computing centers, each with distinct needs [10][11] - Guangbenwei is developing a hybrid optical-electrical computing card that is compatible with existing standards, aiming to provide high energy efficiency and algorithm flexibility [10] Future Prospects - The company has established partnerships with leading internet firms, GPU manufacturers, and intelligent computing centers for application validation and is working on advanced packaging techniques [13] - Guangbenwei has secured multiple rounds of financing to support its growth and product development, indicating strong investor confidence in its potential [13][14]
存算一体瓶颈,中国团队实现突破
半导体芯闻· 2025-07-02 10:21
Core Viewpoint - The rapid development of artificial intelligence (AI) presents new challenges for chip computing power, particularly the "memory wall" issue, which arises from the limitations of the von Neumann architecture widely used in processors [1][3]. Group 1: Memory Wall Problem - The von Neumann architecture simplifies hardware design by storing data and instructions in the same memory, but it limits CPU execution capabilities due to sequential instruction processing [3]. - The performance of storage has not kept pace with CPU advancements, leading to significant delays as CPUs wait for memory read/write operations, thus degrading overall system performance [3][4]. Group 2: Processing-In-Memory (PIM) Technology - PIM, or Compute-in-Memory, is an emerging non-von Neumann computing paradigm aimed at addressing the "memory wall" problem by executing computations within memory, reducing data transfer time and energy costs [5][6]. - The development of PIM technology has evolved through various stages since the 1990s, with significant contributions from both academic institutions and companies like Samsung, SK Hynix, and Micron [6][8]. Group 3: Current PIM Technologies - Mainstream PIM technologies include digital PIM (SRAM/DRAM), analog PIM (RRAM, PCM), and hybrid PIM, each with distinct advantages and challenges [8]. - Companies and research institutions have been actively developing PIM prototypes since 2017, with notable advancements in traditional storage technologies [8][9]. Group 4: Sorting Challenges in AI - Sorting is a critical and time-consuming operation in AI systems, affecting applications in natural language processing, information retrieval, and intelligent decision-making [10][11]. - The complexity of sorting operations, particularly in dynamic environments, poses significant challenges for traditional computing architectures, leading to high time and power consumption [10][11]. Group 5: Breakthrough in Sorting Hardware Architecture - A team from Peking University has achieved a breakthrough in efficient sorting hardware architecture based on PIM technology, addressing the inefficiencies of traditional architectures in handling complex nonlinear sorting tasks [13][14]. - The new architecture reportedly enhances sorting speed by over 15 times and improves area efficiency by more than 32 times, with power consumption reduced to one-tenth of traditional CPU or GPU processors [15][17]. Group 6: Implications and Future Applications - This breakthrough is expected to support a wide range of AI applications, including intelligent driving, smart cities, and edge AI devices, providing a robust foundation for next-generation AI technologies [16][17]. - The successful implementation of this sorting architecture signifies a shift from application-specific solutions to broader, general-purpose computing capabilities within PIM systems [15][16].