Workflow
iOS 19还没来,我提前在iPhone上体验到了苹果最新的AI
AppleApple(US:AAPL) Hu Xiu·2025-05-15 12:04

Core Viewpoint - Apple has quietly released a new visual language model called FastVLM, which shows potential for local execution on devices like iPhone, iPad, and Mac, indicating a shift towards integrating AI deeply into their ecosystem [3][10][60]. Group 1: FastVLM Model Overview - FastVLM is a set of visual language models that can run locally on Apple devices, with three parameter sizes: 0.5B, 1.5B, and 7B [10]. - The model demonstrates impressive performance, with a Time To First Token (TTFT) of 1211 milliseconds for the 1.5B model, indicating a smooth user experience [14]. - FastVLM can recognize common objects and scenes effectively, although it has limitations in Chinese text recognition accuracy [19][35]. Group 2: Technical Innovations - FastVLM is built on Apple's self-developed AI framework MLX and utilizes a new visual encoding backbone called FastViT-HD, which optimizes performance under limited computational power [46][49]. - The model's design allows it to output fewer high-quality visual tokens directly, enhancing inference speed and reducing resource consumption [52][53]. - FastVLM achieves competitive performance with significantly less training data compared to other models, demonstrating efficiency in model training [58]. Group 3: Strategic Implications - The development of FastVLM aligns with Apple's ambition to embed AI as a core component of their products, rather than just an added feature [63][64]. - There are indications that FastVLM may be integrated into future Apple smart glasses, which are expected to be AI-first devices [60][61]. - Apple's approach emphasizes a hardware-defined software strategy, aiming to create a seamless integration of AI across its ecosystem, including iPhones, iPads, and Macs [65][78].