实测拿215项SOTA的Qwen3.5-Omni:摄像头一开,AI给我现场讲论文、撸代码
量子位·2026-03-31 06:43

Core Viewpoint - The article discusses the launch of Qwen3.5-Omni, highlighting its advanced capabilities in multimodal understanding and real-time interaction, which significantly enhance user experience in AI communication [5][51]. Group 1: Product Features - Qwen3.5-Omni achieves true "multimodal" capabilities, seamlessly understanding text, images, audio, and video inputs, and generating detailed scripts with timestamps [5][51]. - It offers three sizes: Plus, Flash, and Light, supporting 256K context and recognizing 113 languages, capable of processing 10 hours of audio or 1 hour of video [6]. - The model has demonstrated strong performance in benchmarks, achieving 215 state-of-the-art (SOTA) results, competing closely with Gemini 3.1 Pro [7][44]. Group 2: Performance Metrics - In audio understanding, Qwen3.5-Omni-Plus scored 84.6 in DailyOmni, surpassing Gemini 3.1 Pro's score of 81.4 [46]. - For visual understanding, it scored 62.8 in WorldSense, while Gemini 3.1 Pro scored 65.5, indicating competitive performance [46]. - The model excels in dialogue and audio recognition, with Qwen3.5-Omni-Plus achieving 93.1 in VoiceBench, outperforming Gemini 3.1 Pro's 88.9 [47]. Group 3: Interaction Capabilities - Qwen3.5-Omni features "vibe coding," allowing it to generate Python code or frontend prototypes during real-time video calls [10][30]. - It supports semantic interruption, enabling users to ask questions or change topics without disrupting the flow of conversation [42]. - The model's architecture allows for real-time processing and generation, making interactions feel more natural and human-like [66][68]. Group 4: Technical Improvements - The model introduces ARIA technology for improved speech stability and naturalness, addressing previous issues of inconsistency in AI speech [64][65]. - It utilizes a hybrid attention mechanism for enhanced efficiency and performance in processing multimodal inputs [55][56]. - The architecture combines a "Thinker" for understanding inputs and a "Talker" for generating speech, allowing for simultaneous processing and output [53][59].

实测拿215项SOTA的Qwen3.5-Omni:摄像头一开,AI给我现场讲论文、撸代码 - Reportify