Core Insights - North American cloud service providers (CSPs) are significantly increasing investments in AI infrastructure, leading to a projected annual growth rate of over 28% in global AI server shipments by 2026 [1] - The demand for AI inference services is driving a replacement and expansion cycle for general servers, with an expected annual growth rate of 12.8% in global server shipments (including AI servers) by 2026 [1] Group 1: Market Trends - From 2024 to 2025, the server market will focus on training advanced large language models (LLMs) using AI servers equipped with GPUs and HBM for parallel computing [3] - Starting in the second half of 2025, the development of AI inference services such as AI Agents, LLaMA model applications, and Copilot upgrades will prompt CSPs to shift towards monetization and profit models [3] - The total capital expenditure growth rate for the five major North American CSPs (Google, AWS, Meta, Microsoft, Oracle) is projected to reach 40% in 2026, driven by large-scale infrastructure investments and the replacement of general servers purchased during the 2019-2021 cloud investment boom [3] Group 2: AI Server Market Dynamics - The 2026 AI server market will be primarily driven by North American CSPs, government sovereign cloud projects, and large CSPs accelerating their own ASIC development and edge AI inference solutions [4] - GPUs are expected to account for 69.7% of AI chip usage, with NVIDIA's GB300 models becoming the mainstream for shipments, while VR200 will gradually ramp up in the second half of the year [4] Group 3: ASIC Development - The share of ASIC AI servers in shipments is expected to rise to 27.8% by 2026, the highest since 2023, with growth rates surpassing those of GPU AI servers [6] - Google is leading the investment in self-developed ASICs, with its TPU not only serving Google Cloud Platform infrastructure but also being sold to external companies like Anthropic [6]
TrendForce集邦咨询:预估2026年全球AI服务器出货同比增逾28%