Workflow
字节开源多模态智能体UI-TARS-1.5 重点强化高阶推理能力
Sou Hu Cai Jing·2025-04-19 06:27

Core Insights - ByteDance's Seed Lab has officially released and open-sourced the next-generation multimodal intelligent agent UI-TARS-1.5, which significantly enhances high-level reasoning capabilities compared to its predecessor [1][2] Performance Metrics - UI-TARS-1.5 outperforms previous models in various benchmarks, achieving scores of 42.5 in OSworld (100 steps), 42.1 in Windows Agent Arena (50 steps), 84.8 in WebVoyager, and 75.8 in Online-Mind2web [2] - The model shows a notable improvement in Android World with a score of 64.2, compared to previous models [2] Technological Advancements - The model's capabilities are derived from four key technological dimensions: enhanced visual perception, System 2 reasoning mechanism, unified action modeling, and self-evolving training paradigm [3] - Enhanced visual perception allows the model to understand interface elements deeply, providing a reliable information foundation for decision-making [3] - The System 2 reasoning mechanism enables multi-step planning and decision-making, mimicking human thought processes [3] - Unified action modeling improves action control and execution precision through a standardized action space [3] - The self-evolving training paradigm allows the model to learn from mistakes and adapt to complex task environments [3] Practical Applications - UI-TARS-1.5 functions as a practical "digital assistant," capable of operating computers and systems, controlling browsers, and completing complex interactive tasks [4]