Core Insights - The article highlights the launch of the latest open-source multimodal world model, Emu3.5, developed by the Beijing Academy of Artificial Intelligence (BAAI), which excels in tasks involving images, text, and videos, showcasing high precision in operations like erasing handwriting [1][6][9]. Group 1: Model Capabilities - Emu3.5 demonstrates advanced capabilities in generating coherent and logical content, particularly in simulating dynamic physical worlds, allowing users to experience virtual environments from a first-person perspective [6][12]. - The model can perform complex image editing and generate visual narratives, maintaining consistency and style throughout the process, which is crucial for long-term creative tasks [15][17]. - Emu3.5's ability to understand long sequences and spatial consistency enables it to execute tasks like organizing a desktop through step-by-step instructions [12][22]. Group 2: Technical Innovations - The model is built on a 34 billion parameter architecture using a standard Decoder-only Transformer framework, unifying various tasks into a Next-State Prediction task [17][25]. - Emu3.5 has been pre-trained on over 10 trillion tokens of multimodal data, primarily from internet videos, allowing it to learn temporal continuity and causal relationships effectively [18][25]. - The introduction of the Discrete Diffusion Adaptation (DiDA) technology enhances image generation speed by nearly 20 times without compromising performance [26]. Group 3: Open Source Initiative - The decision to open-source Emu3.5 allows global developers and researchers to leverage a model that understands physics and logic, facilitating the creation of more realistic videos and intelligent agents across various industries [27][29].
世界模型有了开源基座Emu3.5,拿下多模态SOTA,性能超越Nano Banana
3 6 Ke·2025-10-30 11:56