Workflow
低延迟处理器
icon
Search documents
超百亿美元!OpenAI签下AI芯片大单
新华网财经· 2026-01-16 03:34
Core Viewpoint - OpenAI and Cerebras are collaborating to deploy a 750 MW wafer-scale system, which will become the world's largest high-speed AI inference platform by 2028, with a project value exceeding $10 billion [1]. Group 1: Collaboration and Market Demand - The partnership between OpenAI and Cerebras signifies a strong market demand for inference computing power and highlights the increasing importance of inference speed among tech giants [1]. - Cerebras, founded in 2015, aims to create the fastest AI inference and training platform, with its CS-2 and CS-3 systems already applied in various fields such as medical research and cryptography [4]. Group 2: Technological Advancements - Cerebras' unique system integrates massive computing power, memory, and bandwidth into a single giant chip, eliminating traditional hardware bottlenecks that limit inference speed [4]. - The response speed of large language models based on Cerebras technology can be up to 15 times faster than those based on GPU systems for code and voice chat tasks [4]. Group 3: Industry Trends - The tech industry's history shows that speed has played a crucial role in technology adoption, with significant advancements in processing frequency and internet connectivity driving the growth of personal computing and modern internet [5]. - Low-latency inference solutions provide faster response times and more natural interactions, enhancing productivity in the AI-driven market [5]. Group 4: Competitive Landscape - In December 2025, AI chip startup Groq announced a non-exclusive licensing agreement with NVIDIA, valued at $20 billion, marking NVIDIA's largest transaction to date [5]. - NVIDIA plans to integrate Groq's low-latency processors into its AI factory architecture to support a broader range of AI inference and real-time workloads [6].
超百亿美元!OpenAI签下AI芯片大单
当地时间1月14日,OpenAI与美国AI芯片初创公司Cerebras宣布,将部署750兆瓦的Cerebras晶圆级系统。该合作将于2026年 起分阶段落地,并于2028年完成,建成后将成为全球规模最大的高速AI推理平台。据美国消费者新闻与商业频道 (CNBC)报道,该项合作的价值超过100亿美元。 Cerebras联合创始人兼首席执行官安德鲁·费尔德曼(Andrew Feldman)表示,与OpenAI合作,意味着将全球领先的AI模型 引入全球最快的AI处理器。实时推理将彻底变革AI领域,开启构建和交互AI模型的全新方式。 据悉,Cerebras系统的独特之处在于,其将海量计算能力、内存和带宽集成到单个巨型芯片上,从而消除了传统硬件上制 约推理速度的瓶颈。在代码及语音聊天任务上,基于Cerebras的大语言模型所给出的响应速度比基于GPU的系统快高达15 倍。 通常来说,重点面向逻辑推理的模型往往需要较长时间"思考"后才能生成回应。 但回溯科技行业的发展历程,"速度"在推动技术普及上发挥了重要作用。如果运行频率没有出现从千赫兹到兆赫兹再到吉 赫兹的飞跃,就不会有个人电脑产业;同样,如果没有从拨号上网到宽带网 ...
黄仁勋急疯,全球为缺电抓狂,中国却“电力自由”,底气藏不住了
Sou Hu Cai Jing· 2025-12-31 16:47
西方电力捉襟见肘 现在人工智能发展得飞快,可这也带来大麻烦,尤其是电力供应跟不上。英伟达的黄仁勋今年在乔罗根 的节目里直言,AI扩展太猛,电力成了新瓶颈。要是不解决,芯片再厉害也白搭。他还预测,AI计算 需求2025年已经暴涨,以前以为芯片是核心,现在电力更关键。 要知道,英伟达自家处理器集群到2027年一年可能吃掉150到200吉瓦电,这相当于一个国家全年用电 量。黄仁勋还警告,美国数据中心建得慢,电力容量不够,可能让中国在AI赛道上超车。他甚至说, 未来科技公司得自己搞小型核反应堆来发电,不然AI就卡壳了。这话不是随便说说,英伟达员工邮件 里也提了整合低延迟处理器,但电力问题始终绕不开。 发展中国家更惨,像乍得、马拉维,农村地区电力覆盖率低,基础设施跟不上。 中国电网游刃有余 中国这边,电力供应倒是不慌,底气来自于多年积累的基础设施和新能源布局。早些年,我们也尝过缺 电的苦,90年代广东工厂开三停四,拉闸限电是家常便饭。但从2000年启动西电东送工程开始,情况就 变了。这项目把西部能源富集区电力输到东部需求大户,通过特高压技术实现高效传输。特高压就是电 压超高的传输方式,交流1000千伏,直流±800千伏 ...
未直接并购避监管?英伟达200亿美元引进Groq推论技术等
Ju Chao Zi Xun· 2025-12-25 13:13
Core Insights - Nvidia has agreed to acquire certain assets of startup Groq for $20 billion, marking the largest asset and technology transaction in Nvidia's history [1] - The deal includes a non-exclusive technology licensing agreement for Groq's inference chip technology and the recruitment of key executives, including Groq's founder Jonathan Ross [1][2] - Groq recently completed a funding round with a valuation of approximately $6.9 billion, raising $750 million from investors including Blackstone and Samsung Electronics [1] Group 1 - The transaction was disclosed by Alex Davis, CEO of the venture capital firm that led Groq's recent funding round, indicating that negotiations were relatively brief [1] - Nvidia's CEO Jensen Huang stated that the integration of Groq's low-latency processors will support a broader range of real-time and inference workloads within Nvidia's AI factory architecture [2] - Groq will continue to operate as an independent company post-transaction, with its CFO Simon Edwards taking over as CEO, and its cloud service GroqCloud is not included in the deal [2]