Core Viewpoint - AI inference is not only a technological revolution but also a highly profitable business that can be precisely calculated [1][2]. Group 1: Profitability Analysis - Morgan Stanley's report reveals that a standard "AI inference factory" has an average profit margin exceeding 50%, with Nvidia's GB200 leading at nearly 78% [2][6]. - Google's TPU v6e pod follows closely with a profit margin of 74.9%, demonstrating the economic efficiency of top cloud providers through hardware and software optimization [10]. - AWS's Trn2 UltraServer and Huawei's Ascend CloudMatrix 384 platform achieve profit margins of 62.5% and 47.9%, respectively [11]. - In contrast, AMD's platforms, MI300X and MI355X, show significant losses with profit margins of -28.2% and -64.0%, attributed to high costs and low output efficiency [12]. Group 2: 100MW AI Factory Model - Morgan Stanley introduces the "100MW AI factory model," which standardizes the evaluation of different AI solutions based on a typical medium-sized data center's power consumption [15]. - The model calculates total cost of ownership (TCO) for a 100MW AI factory, estimating annual TCO between $330 million and $807 million [16]. - Revenue is directly linked to token output, with a fair price set at $0.2 per million tokens, considering a 70% utilization rate for realistic revenue predictions [16]. Group 3: Future Landscape and Strategic Competition - The report highlights that the future AI landscape will focus on building technological ecosystems and product roadmaps [19]. - A battle over "connection standards" is emerging among non-Nvidia players, with AMD advocating for UALink and Broadcom supporting a more open Ethernet approach [19]. - Nvidia is solidifying its lead with a clear roadmap for its next-generation platform "Rubin," expected to enter mass production in Q2 2026 [19].
大摩建模“AI推理工厂”:无论是英伟达还是华为芯片,都能盈利,平均利润率超50%