字节视觉-语言多模态大模型Seed VLM技术报告首次公开
news flash·2025-05-13 06:16
Core Insights - ByteDance's Seed team has released the latest visual-language multimodal model, Seed1.5-VL, which demonstrates enhanced general multimodal understanding and reasoning capabilities [1] - The model significantly reduces reasoning costs and achieves state-of-the-art (SOTA) performance in 38 out of 60 public evaluation benchmarks [1] - Seed1.5-VL is now available for user experience through an API on the Volcano Engine [1]