Core Insights - Apple is deepening its "vertical integration" strategy by developing its first self-designed AI server chip, codenamed "Baltra," in collaboration with Broadcom, aiming to reduce reliance on Nvidia chips [1][2] - The "Baltra" chip is specifically designed for "AI inference," focusing on executing tasks rather than training large-scale AI models, with Apple opting to rent Google's custom Gemini model for $1 billion annually [1] - The architecture of "Baltra" will differ significantly from traditional training chips, emphasizing low latency and high concurrent throughput, with a focus on optimizing low-precision mathematical operations to reduce energy consumption and enhance user response speed [2] Company and Industry Summary - The collaboration with Broadcom is crucial for overcoming core network transmission technology challenges, with the chip expected to be operational by 2027 [1] - The chip is likely to utilize TSMC's advanced 3nm "N3E" process, with design work anticipated to be completed within the next 12 months [2] - The strategic focus on inference rather than training aligns with industry trends towards optimizing AI performance for user-facing applications [2]
苹果携手博通研发AI服务器芯片Baltra,2027年投入使用