Workflow
内存墙
icon
Search documents
一家水下AI芯片公司完成10亿元融资,瞄准大模型推理
暗涌Waves· 2026-02-13 00:57
Core Viewpoint - The article discusses the rapid development and funding of a 3D AI chip company, 算苗科技 (Suanmiao Technology), which has completed two rounds of financing totaling nearly 1 billion RMB, aimed at developing domestically produced 3D computing chips for AI applications [3][10]. Group 1: Company Overview - 算苗科技 focuses on the research and development of 3D computing chips, with its core product being a customized chip for AI model inference [4]. - The company aims to address the "memory wall" issue that limits AI model computation, as current AI chips face significant inefficiencies due to memory bandwidth constraints [4][5]. - 算苗科技's A4 chip has demonstrated a throughput of 1.26 to 2.19 times that of NVIDIA's H200 in inference tasks on major open-source models [5]. Group 2: Funding and Market Position - The recent funding rounds were led by prominent investors, including Source Code Capital and Shixi Capital, indicating strong market interest and support for the company's vision [3][10]. - The company is positioned to leverage its expertise in 3D IC technology to create a competitive edge in the AI chip market, which is expected to grow significantly [10][19]. Group 3: Technological Innovation - 算苗科技 utilizes a 3D stacked architecture that allows for significantly higher memory bandwidth (up to 32 TB/s), which is crucial for AI model inference [4][13]. - The company’s approach contrasts with traditional GPU architectures, focusing on specialized ASIC designs that optimize performance for specific tasks rather than general-purpose computing [14][15]. Group 4: Strategic Focus - The company has chosen to concentrate on AI model inference rather than training, as it anticipates that 90% of future AI computing demand will be for inference tasks [15][18]. - 算苗科技 believes that the future of AI computing lies in architectural innovation, particularly through 3D stacking and ASIC optimization, which aligns with the growing demand for efficient computing solutions [28][29].
一家光芯片公司,获2.2亿美元融资
半导体芯闻· 2026-02-12 10:37
据英国《金融时报》今日报道,本轮融资由比利时风投机构Hummingbird Ventures领投,交易完 成 后 , 总 部 位 于 英 国 的 Olix Computing 估 值 超 10 亿 美 元 。 该 公 司 此 前 曾 从 Plural 、 Vertex Ventures、LocalGlobe 及 Entrepreneurs First 等机构获得未披露金额的融资。 如果您希望可以时常见面,欢迎标星收藏哦~ 英国初创企业Olix Computing正在研发集成光学组件的人工智能芯片,该公司已完成2.2 亿美元融 资。 Olix 的芯片针对AI 推理进行优化,推理是指 AI 模型完成训练后,在实际业务环境中运行的过 程。目前尚不清楚该处理器集成了哪些光学组件,以及具体应用方式。不过,该公司官网的一篇博 客文章提到,其芯片采用"创新型内存与互联架构"。这表明 Olix 正在利用光子组件构建互联模 块,即处理器中负责在电路之间传输数据的部分。 目前已有多家初创企业在研发光子互联技术。其中融资规模领先的Ayar Labs已开发出一款光学中 介 层 ( interposer ) , 可 用 于 打 造 面 ...
光子AI芯片初创公司Olix获得2.2亿美元投资
Sou Hu Cai Jing· 2026-02-12 09:16
英国初创公司Olix Computing Ltd.开发集成光学组件的人工智能芯片,近日宣布获得2.2亿美元融资。 据Olix介绍,其芯片设计通过不使用HBM来解决这一挑战。该处理器仅使用SRAM存储数据,这是一种 速度显著更快的内存类型。 HBM内存单元由一个晶体管和一个微型电池(电容器)组成,而SRAM使用六个晶体管的更复杂设 计。SRAM性能优势的另一个因素是它通常直接集成到AI芯片中,而HBM内存作为独立模块实现,这 意味着SRAM更接近主芯片的晶体管,从而减少数据传输时间。 初创公司Cerebras Systems Inc.在设计其晶圆级AI加速器时也优先考虑了SRAM。该芯片包含44GB的 SRAM,使许多AI模型无需使用HBM即可运行。Olix声称其光子技术在交互性和延迟方面优于"纯硅 SRAM架构"。 Olix的芯片被称为OLIX光学张量处理单元,简称OTPU。张量是AI模型用来保存信息的数学对象,许多 AI芯片都包含专门优化来处理此类对象的电路。 据《金融时报》报道,本轮投资由比利时风险投资公司Hummingbird Ventures领投,使这家英国公司的 估值超过10亿美元。此前,Olix还 ...
DRAM危机,短期无解
半导体行业观察· 2026-02-11 01:27
公众号记得加星标⭐️,第一时间看推送不会错过。 如果感觉如今科技领域的一切都与人工智能息息相关,那是因为事实的确如此。而计算机内存市场更 是如此。用于为人工智能数据中心中的GPU和其他加速器供电的DRAM内存需求巨大,利润丰厚,以 至于挤占了其他用途的内存供应,导致价格飙升。据Counterpoint Research的数据显示,本季度 DRAM价格已上涨80%至90%。 最大的人工智能硬件公司表示,他们已经确保了芯片供应到 2028 年,但这让其他所有人——个人电 脑制造商、消费电子产品制造商以及其他所有需要临时存储十亿比特数据的设备制造商——都不得不 争先恐后地应对供应短缺和价格上涨的问题。 电子行业是如何陷入如今这般困境的?更重要的是,它又该如何摆脱困境?IEEE Spectrum采访了经 济学家和存储器专家,请他们对此进行解释。他们认为,如今的局面是 DRAM 行业历史上长期存在 的繁荣与衰退周期,以及规模空前的 AI 硬件基础设施建设相互碰撞的结果。而且,除非 AI 领域出 现重大崩盘,否则新增产能和新技术需要数年时间才能使供应与需求相匹配。即便如此,价格也可能 依然居高不下。 要了解事情的来龙去脉 ...
这种芯片将突破内存壁垒
半导体行业观察· 2026-02-10 01:14
人们正在寻找能够突破人工智能长期存在的"内存墙"的方法——即使是快速模型也会因处理器和内存 之间数据传输所需的时间和能量而运行缓慢。电阻式随机存取存储器(RRAM)可以通过允许计算在 内存本身进行来绕过这一障碍。然而,大多数类型的非易失性存储器都过于不稳定和笨重,无法用于 此目的。 幸运的是,可能已经有了解决方案。在12月的IEEE国际电子器件会议(IEDM)上,加州大学圣地亚 哥分校的研究人员展示了他们可以在一种全新的RRAM上运行学习算法。 "我们实际上重新设计了 RRAM,彻底重新思考了它的开关方式,"领导这项工作的加州大学圣地亚 哥分校电气工程师杜伊古·库祖姆 (Duygu Kuzum)说。 公众号记得加星标⭐️,第一时间看推送不会错过。 RRAM 将数据存储为对电流流动的电阻值。神经网络中的关键数字运算——将数字数组相乘然后将 结果相加——可以通过模拟方式轻松实现:只需让电流流过 RRAM 单元阵列,连接它们的输出,然 后测量产生的电流即可。 传统上,RRAM 通过在介电材料的高电阻环境中形成低电阻细丝来存储数据。形成这些细丝通常需 要过高的电压,超出了标准CMOS 工艺的承受范围,从而阻碍了其在处 ...
中国推理芯片突围与成本革命:破“内存墙”、兼容CUDA
Core Insights - The article discusses the shift in the global AI computing power focus from training to inference, indicating a competitive landscape for cost-effective and energy-efficient chips [1][2] - The consensus in the industry is that inference chips will dominate AI evolution in the next five to ten years, with companies like Google and Nvidia leading the charge [1][3] - CloudWalk Technology has announced its strategic focus on AI inference chips, aiming to significantly reduce the cost of processing tokens, which are becoming a core productivity driver in the AI landscape [2][3] Industry Trends - The demand has shifted from relying on high-performance GPUs to a pressing need for high-cost performance inference chips [2] - The past year has seen a dramatic increase in the computational requirements for large models, with token processing needs growing hundreds of times, highlighting the importance of inference over training [2][3] - Nvidia's strategic acquisition of Groq's core assets for $20 billion reflects the growing importance of inference chips, with Groq's valuation skyrocketing from $7 billion to $20 billion in just four months [3] Company Strategy - CloudWalk Technology's CEO, Chen Ning, emphasizes the goal of reducing the cost of processing one million tokens by 100 times, aiming for a transformative impact on industrial productivity by 2030 [3][4] - The company is developing a new processor architecture, GPNPU, designed to optimize inference for large models while addressing cost, efficiency, and deployment challenges [5][6] - The GPNPU architecture aims to maintain compatibility with existing CUDA programs, lowering the barrier for integration into production systems [5][6] Product Development - CloudWalk Technology plans to launch the DeepVerse 100, 200, and 300 series chips over the next five years, targeting major clients across various industries [6] - The company is focusing on modular chip design through a "power building block" approach, allowing for scalable and flexible computing solutions [6] - The company has established a strong domestic production capacity, ensuring supply chain security for large-scale chip production and delivery [6]
100根内存条换一套房,AI疯狂吞噬全球内存,普通人电脑快买不起了
3 6 Ke· 2026-01-20 07:22
Core Insights - The tech industry is facing a significant crisis due to a "memory wall," which is limiting the growth of AI despite high expectations for computational power [1][2][10] - The demand for DRAM (Dynamic Random Access Memory) is expected to surge, with prices projected to increase by 88% in 2026, driven by the insatiable needs of AI data centers [2][8] - The current memory crisis is causing a supply shortage for consumer electronics, leading to higher prices for devices like computers and smartphones [5][10] Group 1: Memory Crisis and Price Surge - The price of DDR5 memory has increased by approximately 307% since September 2025, with high-capacity server memory modules reaching prices of 400,000 yuan for 100 units [6][8] - Citibank has revised its forecast for DRAM prices, predicting an increase of 88% in 2026, up from a previous estimate of 53% [2][8] - The demand from AI giants like OpenAI and Google is consuming a significant portion of memory production, leading to a scarcity in the consumer market [5][10] Group 2: Impact on Consumers and Market Dynamics - Consumers are experiencing higher prices for new computers and smartphones, with the market supply for consumer-grade memory dwindling [5][10] - PC vendors are prioritizing supply for large OEMs, resulting in reduced availability for third-party module manufacturers [10] - The ongoing memory crisis is perceived as a "resource tax" that consumers are forced to pay for the advancement of AI technology [10] Group 3: Technological Implications and Future Outlook - The growth of AI models is outpacing the advancements in memory bandwidth, creating a bottleneck that could hinder further AI development [13][14] - Innovations such as High Bandwidth Memory (HBM) and new architectures like CXL and PIM are being explored to overcome the memory wall [18][19] - The trend indicates that the era of affordable and abundant memory is coming to an end, with implications for both consumers and AI companies [19][20]
存储猛拉,AI存力超级周期到底有多神?
3 6 Ke· 2026-01-06 12:19
Core Insights - The storage industry is experiencing a significant upcycle driven by AI demand, extending from HBM to traditional storage sectors, with Micron's gross margin reaching a historical high of 66-68% for the next quarter, indicating a stronger cycle than previous ones [1][3]. Group 1: AI Demand and Storage Market Dynamics - The price increase of storage products reflects the supply-demand relationship in the market, primarily driven by AI server demand [3]. - The current AI storage cycle is characterized by a shift in focus from training to inference, leading to differentiated demands for "low latency, high capacity, and high bandwidth" storage [3][14]. - The three major manufacturers (Samsung, SK Hynix, Micron) are prioritizing capital expenditures towards HBM and DRAM, resulting in structural supply-demand imbalances and significant price increases [3][6]. Group 2: Role of Different Storage Types in AI Servers - HBM serves as the "performance ceiling" for AI servers, being a high-bandwidth, high-power product that directly impacts the model scale and response speed [11]. - DRAM (DDR5) acts as a data exchange hub, connecting HBM and NAND, and is crucial for handling concurrent tasks in AI servers [12]. - NAND (SSD) functions as a fast persistent layer for frequently accessed data, while HDD serves as a low-cost container for large volumes of cold data [12][14]. Group 3: Addressing the "Memory Wall" Challenge - The "memory wall" bottleneck arises from the disparity between computing speed and data transfer speed, leading to high GPU idle rates [5][16]. - Solutions to this issue include upgrading HBM to 16-Hi stacks to enhance bandwidth and implementing 3D stacked SRAM to reduce latency [18][19]. - The integration of computing capabilities within storage (compute-in-memory) is anticipated to be a long-term solution to the "memory wall" problem [21]. Group 4: HBM Market Supply and Demand - HBM demand is closely tied to AI chip shipments, with expectations for HBM supply to increase by over 60% by 2026 due to significant capital investments from the three major manufacturers [6][24]. - The combined monthly HBM production capacity of the three manufacturers is projected to rise from approximately 390,000 wafers to 510,000 wafers by the end of 2026, translating to an estimated supply of 41.9 billion GB of HBM [29][34]. - The HBM market is expected to be in a "tight balance" state in 2026, with demand estimated at around 42 billion GB, indicating a competitive landscape among manufacturers [39][40].
美国制造一颗真正的3D芯片
半导体行业观察· 2025-12-13 01:08
Core Insights - A collaborative team has developed the first monolithic 3D chip at a U.S. foundry, achieving unprecedented vertical wiring density and speed improvements [2][3] - This innovation is expected to usher in a new era of AI hardware and domestic semiconductor innovation, with the potential for a 1000-fold increase in hardware performance needed for future AI systems [3][7] Group 1: Performance and Design - The new 3D chip's performance is approximately an order of magnitude higher than traditional 2D chips, addressing the long-standing limitations of flat designs [2][3] - Early hardware tests indicate that the prototype chip outperforms similar 2D chips by about four times, with simulations suggesting up to a 12-fold performance increase for future versions [7] - The design allows for a significant enhancement in energy-delay product (EDP), potentially improving it by 100 to 1000 times, which balances speed and energy efficiency [7] Group 2: Technical Challenges and Solutions - Traditional 2D chips face a "memory wall" bottleneck due to the slow data transfer speeds compared to processing speeds, limiting overall system performance [4][5] - The new chip overcomes these challenges by vertically integrating memory and computation, allowing for faster data transmission and higher density connections [5][6] - Unlike previous attempts that relied on stacking separate chips, this new approach uses a continuous process to stack layers directly, enhancing connection density and manufacturability [6] Group 3: Implications for the Semiconductor Industry - The successful production of this 3D chip in a domestic foundry signifies a major step for U.S. semiconductor innovation, indicating that advanced architectures can be commercially viable [6][7] - The transition to vertical monolithic 3D integration will require a new generation of engineers skilled in these technologies, fostering a new wave of innovation in the semiconductor field [7][8] - The breakthrough not only enhances performance but also positions the U.S. to lead in the future of AI hardware development [8]
传迈威尔科技(MRVL.US)拟斥资50亿美元收购Celestial AI 押注光子互联破局“...
Xin Lang Cai Jing· 2025-12-02 06:57
Group 1 - Marvell Technology is in advanced talks to acquire Celestial AI for a deal potentially exceeding $5 billion, including cash and stock [1] - The acquisition is expected to enhance Marvell's product portfolio and highlight the strong demand for computing power in the market [1] - Celestial AI has raised a total of $515 million, with $250 million coming from a venture capital round supported by an AMD subsidiary [1][2] Group 2 - Celestial AI is focused on developing a photonic interconnect platform called Photonic Fabric to address the "memory wall" crisis in AI computing architectures [2] - The memory wall has become a significant barrier to system performance as the speed mismatch between computing units and memory leads to inefficiencies, especially with large AI models [2] - The acquisition of Celestial AI would provide Marvell with a strategic advantage in the evolving AI server market, particularly if photonic interconnect technology becomes a standard [3]