Core Insights - The development of multimodal large models positions physical AI as a key battleground among tech giants, with Alibaba Cloud recently launching a multimodal interaction development kit at the Tongyi Intelligent Hardware Expo in Shenzhen [1] Group 1: Development and Trends - Alibaba Cloud's Tongyi large model business manager predicts significant growth in smart hardware by 2026, with new categories emerging beyond traditional devices like smartphones and cars [2] - The transition of models from "cloud" to "on-device" is expected to enhance usability and foster ecosystem prosperity, potentially leading to closed-loop business models [2] - Current AI hardware faces multiple barriers to widespread adoption, including high costs, privacy concerns, and outdated development paradigms [2] Group 2: Features of the Development Kit - The newly launched multimodal interaction development kit offers a low barrier to entry, fast response times, and a rich array of scenarios for hardware companies and solution providers [2][7] - A notable feature of the kit is the change in billing method from token-based to hardware terminal-based service provision, significantly reducing costs for developers [3] Group 3: Specialized Solutions - Alibaba Cloud showcased solutions for smart wearables, companion robots, and embodied intelligence, featuring a complete interaction chain for AI glasses that includes real-time translation and transcription capabilities [4] - The kit supports various interaction methods, achieving low latency in voice and video interactions, and integrates with Alibaba Cloud's Bailian platform ecosystem [4][7] Group 4: Collaboration and Market Position - Alibaba Cloud is open to collaborating with developers to lower barriers and explore deep partnerships in various verticals, particularly in the smartphone and smart glasses markets [5] - The company is focusing on two main technological routes: GUI and A2A, with the latter showing faster development and better user experience [5] Group 5: Future Developments - The multimodal interaction kit is designed for vertical applications and is compatible with over 30 mainstream terminal chip platforms, aiming for optimized deployment and inference performance [7] - Alibaba Cloud is exploring the VLA (Visual-Language-Action) model for embodied intelligence, which is still in early stages but expected to yield results by late 2026 [8]
阿里云“抢滩”物理AI 构建多模态硬件生态闭环
Zheng Quan Shi Bao Wang·2026-01-12 12:59