Group 1 - OpenAI and Cerebras announced a partnership to deploy a 750 MW wafer-scale system, which will become the world's largest high-speed AI inference platform by 2028, with a project value exceeding $10 billion [1] - Cerebras' CEO Andrew Feldman emphasized that this collaboration will integrate leading AI models with the fastest AI processors, revolutionizing real-time inference in the AI field [1] - Analysts believe this deal highlights the strong market demand for inference computing power and the increasing importance of inference speed among tech giants [1] Group 2 - Founded in 2015, Cerebras aims to create the fastest AI inference and training platform, with its CS-2 and CS-3 systems already applied in various fields such as medical research and cryptography [4] - Cerebras' unique system integrates massive computing power, memory, and bandwidth into a single giant chip, eliminating traditional hardware bottlenecks, achieving response speeds up to 15 times faster than GPU-based systems for code and voice chat tasks [4] - The tech industry's history shows that speed has been crucial for technology adoption, with low-latency inference solutions providing faster responses and enhancing user interaction, which is vital for the growth of AI-driven markets [4] Group 3 - In December 2025, AI chip startup Groq signed a non-exclusive licensing agreement with NVIDIA, valued at $20 billion, marking NVIDIA's largest deal to date [5] - NVIDIA plans to integrate Groq's low-latency processors into its AI factory architecture to expand its platform for broader AI inference and real-time workloads [5]
超百亿美元!OpenAI签下AI芯片大单