Workflow
对话红帽全球副总裁曹衡康:AI成本下降了 芯片的量一定会起来
Mei Ri Jing Ji Xin Wen·2025-06-14 09:02

Core Viewpoint - The consensus in the industry is that the cost of computing power will eventually decrease, but there is no unified path chosen among data centers, integrated machines, or inference servers [1] Group 1: AI Inference Year - 2023 is considered the year of AI inference, marking the official launch of AI applications that will generate business revenue and internal cost control for enterprises [1] - Red Hat has chosen to adopt the "vLLM" framework, a high-performance large language model inference framework that has become a de facto standard in the open-source community [1] Group 2: Contribution and Market Potential - Contributors from China account for 35% of the contributions to the vLLM community, indicating a strong potential for inference technology to bring enterprise value in China [1] - The company identifies two technical challenges in inference: achieving high-performance inference with minimal hardware and cost, and distributing inference workloads across multiple servers [1] Group 3: Future of Computing Power Costs - Red Hat plans to launch inference servers in 2025, emphasizing that the main advantage is the reduction of computing power costs for enterprises [2] - The company does not produce hardware but focuses on software solutions, aiming to lower the barriers for AI adoption among businesses [2] - As computing costs decrease, the demand for GPU cards is expected to rise significantly, potentially increasing the number of enterprises using AI from 1,000 to 100,000 or even 1 million [2]