Investment Rating - The industry investment rating is "Positive" (maintained) [2] Core Insights - The release of the Rubin CPX inference chip by Nvidia emphasizes cost-effectiveness and is designed specifically for large-scale context AI models, providing 20 PFLOPS of computing power with a memory bandwidth of 2TB/s, significantly reducing memory costs by over 50% by switching from HBM to GDDR7 [5][6] - The introduction of the Rubin CPX chip expands the VR200 server architecture into three versions, which is expected to create new opportunities in the supply chain, particularly increasing demand for PCB and copper cable connectors due to the complexity of interconnections [6][7] Summary by Sections Industry Investment Rating - The report maintains a "Positive" rating for the industry, indicating expectations for the industry to outperform the overall market [2] Nvidia Rubin CPX Chip - The Rubin CPX chip is designed for two critical stages of AI inference: Prefill and Decode, with a focus on maximizing computational throughput while minimizing memory bandwidth waste [5] - The chip features a design that prioritizes computational FLOPS over memory bandwidth, making it suitable for high-demand AI applications [5] Supply Chain Opportunities - The new architecture introduced by the Rubin CPX chip is anticipated to generate additional demand in the supply chain, particularly for PCB and copper cable connectors, as the complexity of interconnections increases [6][7] - Beneficiary companies in the PCB segment include Huadian Co., Shenghong Technology, and others, while copper cable connector beneficiaries include Huafeng Technology and others [7]
推理专用芯片RubinCPX重磅发布,产业链迎来新机遇