Workflow
AI视频生成
icon
Search documents
硬刚Sora2,万相2.6轻松定制角色、控制分镜,普通人也能当导演
机器之心· 2025-12-17 05:28
Core Insights - The article highlights the rapid advancements in video generation technology, particularly focusing on the release of Alibaba's Wan 2.6 model, which significantly enhances user capabilities in video creation and storytelling [1][36]. Group 1: Technological Advancements - OpenAI's Sora 2 introduced a "Cameo" feature that addresses the "character consistency" issue in AI video generation, transforming the process from unpredictable to controllable [1]. - Alibaba's Wan 2.6 model is noted for its comprehensive capabilities, including voice and image synchronization, allowing users to create videos with a high degree of realism and narrative coherence [3][9]. - The new model supports a maximum video generation duration of 15 seconds, which is the highest in the domestic market, and includes a "shot control" feature for professional storytelling [3][4]. Group 2: User Experience and Accessibility - The Wan 2.5 version of the model made video creation accessible on mobile devices, while the 2.6 version further democratizes professional video production, enabling anyone to take on roles like director or actor [2][4]. - Users can create videos with high fidelity in both visual and auditory aspects, showcasing the model's ability to replicate character traits and emotional expressions accurately [11][24]. Group 3: Practical Applications - The model's capabilities extend to generating complete narrative short films, making it suitable for advertising design and short drama production [16]. - The article emphasizes the model's potential in various creative fields, including AI comic production, advertising design, and short video creation, with over ten visual creation capabilities supported [35][36]. Group 4: Conclusion and Future Implications - The release of Wan 2.6 signifies a shift from a mere "lottery" approach in AI video generation to a new phase of precise and controllable cinematic creation [36]. - The technology effectively removes barriers to creativity, allowing users to leverage their imagination as their primary production tool [37].
千问App接入视频生成模型万相2.6
Mei Ri Jing Ji Xin Wen· 2025-12-17 03:30
Core Viewpoint - Alibaba's Qianwen App has launched the latest video generation model, Wanshang 2.6, and made it available for free to all users, introducing the "AI Theater" feature that allows users to co-star with friends or celebrities in AI-generated short films [1] Group 1: Product Features - The "AI Theater" feature is based on the new role-playing functionality of the model, which is currently only matched by OpenAI's Sora2 model in the U.S. [1] - Since its public beta launch on November 17, the Qianwen App has undergone rapid iterations, with a total of 18 updates in one month [1]
商汤发布Seko2.0:已能连贯创作百集短剧,适配寒武纪
Nan Fang Du Shi Bao· 2025-12-17 01:01
Core Insights - AI video generation is transitioning from a "show-off" phase to a "business realization" phase, with a focus on cost reduction and content consistency [2] - The launch of Seko 2.0 by SenseTime emphasizes the importance of multi-episode consistency and significant advancements in domestic computing power adaptation [2][3] Group 1: Cost Reduction - The inference cost of Seko has decreased by approximately 50% due to model distillation, operator optimization, and adaptation to domestic AI chips [2][3] - This reduction in cost is crucial for B-end users, particularly short drama studios, as it directly impacts project profitability [4] Group 2: Technological Advancements - Seko 2.0 introduces technologies like SekoIDX for consistency and SekoTalk for audio-visual synchronization, enabling coherent creation of up to 100 episodes [6] - The integration of domestic hardware with SenseTime's software has allowed for a seamless transition from reliance on imported computing power, fostering a more resilient domestic industry chain [4] Group 3: Content Consistency - The challenge of maintaining character consistency and narrative coherence in multi-episode productions has been a significant barrier for AI video tools [5][6] - The successful application of Seko 2.0 in real-world scenarios, such as the short drama "Wanxin Ji," demonstrates the effectiveness of a hybrid human-AI collaboration model in film production [6]
一键生成电影级短片,阿里发布万相2.6视频模型
Guo Ji Jin Rong Bao· 2025-12-16 06:59
该模型进一步降低了专业视频制作的门槛。例如,用户上传一段个人视频并输入风格提示词,万相 2.6即可自动完成分镜设计、角色演绎与画面配音,生成电影级短片。在广告设计、短剧制作等场景 中,连续提示词可驱动模型生成叙事连贯、镜头语言专业的视频内容。 据记者了解,目前,万相模型家族已支持文生图、图像编辑、文生视频、角色扮演等十余项视觉创 作能力,广泛应用于AI漫剧、广告创意及短视频制作等领域。 12月16日,《国际金融报》记者获悉,阿里发布新一代万相2.6系列模型,该系列模型面向专业影 视制作和图像创作场景进行了全面升级,是国内首个支持角色扮演功能的视频模型。该模型同时支持音 画同步、多镜头生成及声音驱动等功能,是全球功能最全的视频生成模型。目前万相2.6已同步上线阿 里云百炼、万相官网。 万相2.6集成了音画同步、多镜头生成及声音驱动等多项功能,被官方称为全球功能覆盖最全面的 视频生成模型之一。相比今年9月发布的万相2.5,新版本在画质、音效与指令遵循能力上均有显著提 升,单次视频生成时长达到目前国内最高的15秒,并新增角色扮演与分镜控制功能。这些升级使得模型 能够一键生成包含单人、多人或人与物互动的视频内容,并可 ...
阿里发布电影级视频模型万相2.6,人人都能当电影主角
Ge Long Hui· 2025-12-16 06:11
(原标题:阿里发布电影级视频模型万相2.6,人人都能当电影主角) 12月16日,阿里发布新一代万相2.6系列模型,该系列模型面向专业影视制作和图像创作场景进行了全 面升级,全新的万相2.6是国内首个支持角色扮演功能的视频模型。该模型同时支持音画同步、多镜头 生成及声音驱动等功能,是全球功能最全的视频生成模型。据记者了解,万相2.6已同步上线阿里云百 炼、万相官网。 今年9月,阿里率先在国内发布音画同步的视频生成模型万相2.5,极大提升视频创作的效率,在权威大 模型评测集LMArena上,万相图生视频位居国内第一。此次发布的万相2.6进一步提升了画质、音效、 指令遵循等能力,单次视频时长实现国内最高的15s,万相2.6还新增了角色扮演和分镜控制功能,不仅 能一键完成单人、多人、人与物合拍的视频,还能自动实现多镜头切换等专业任务,满足专业影视级场 景需求。 万相2.6是国内首个支持角色扮演功能的模型,能参考输入视频中的角色外观和音色、实现角色扮演, 按照提示词生成单人、多人、人与物合拍的视频 据介绍,通义万相在模型结构上集成了多项创新技术,可对输入参考视频进行多模态联合建模与学习, 参考具有时序信息的主体情绪、姿 ...
阿里发布通义万相2.6系列视频生成模型,上线国内首个角色扮演功能 | 钛快讯
Tai Mei Ti A P P· 2025-12-16 05:22
12月16日消息,阿里发布新一代万相2.6系列模型,该系列模型面向专业影视制作和图像创作场景进行 了全面升级,全新的万相2.6是国内首个支持角色扮演功能的视频模型。该模型同时支持音画同步、多 镜头生成及声音驱动等功能,是目前全球功能最全的视频生成模型。万相2.6已同步上线阿里云百炼、 万相官网。 今年9月,阿里率先在国内发布音画同步的视频生成模型万相2.5,极大提升视频创作的效率,在权威大 模型评测集LMArena上,万相图生视频位居国内第一。 此次发布的万相2.6进一步提升了画质、音效、指令遵循等能力,单次视频时长实现国内最高的15s,万 相2.6新增了角色扮演和分镜控制功能,不仅能一键完成单人、多人、人与物合拍的视频,还能自动实 现多镜头切换等专业任务,满足专业影视级场景需求。 万相2.6是国内首个支持角色扮演功能的模型,能参考输入视频中的角色外观和音色、实现角色扮演, 按照提示词生成单人、多人、人与物合拍的视频 据介绍,通义万相在模型结构上集成了多项创新技术,可对输入参考视频进行多模态联合建模与学习, 参考具有时序信息的主体情绪、姿态和多角度全面视觉特征,同时提取音色、语速等声学特征,在生成 阶段作为参 ...
商汤发布业内首个多剧集生成智能体Seko2.0
Zheng Quan Ri Bao Wang· 2025-12-15 11:43
Group 1 - SenseTime Group Limited officially launched Seko2.0, the first integrated AI video generation tool designed for individual creators and studios in the short drama and comic drama sectors [1] - Seko1.0, launched in July 2025, quickly gained over 100,000 users within two months and surpassed 200,000 creators in less than six months, with 30% being short drama creators and 20% comic drama creators [1] - Seko2.0 features a new UI upgrade for enhanced visual effects, supports multi-episode generation with context management for up to 100 episodes, and includes intelligent scheduling for characters and scenes [1] Group 2 - The creation of AI short dramas and comic dramas involves generating a vast number of frames, with a single 5-second video requiring nearly 100,000 tokens, and generating 10-20 frames could demand 1,000,000 to 2,000,000 tokens [2] - SenseTime introduced Phased DMD distillation technology to significantly reduce the overall cost of multi-episode generation by combining staged distillation with expert mixture models [2] - The company has open-sourced LightX2V, the first real-time video generation inference framework, which can generate a 5-second video in under 5 seconds on consumer-grade graphics cards, with over 3.5 million downloads [2] Group 3 - SenseTime plans to offer domestic solutions for creators, aiming to initiate a "low-cost" era in AI video creation with exceptional cost-performance [3]
AI视频进入“连续叙事”时代:“一致性”成商业化发展瓶颈
Core Insights - The AI video generation industry is undergoing a paradigm shift, moving from a focus on single-frame realism to multi-episode consistency and sustainable creation [1] - The key challenge for the industry by the second half of 2025 will be whether content can be sustained over multiple episodes, emphasizing character consistency and world coherence [1] Group 1: Industry Trends - AI video generation technology has transitioned from "can it move" to "does it move like it should," highlighting the importance of continuity in storytelling [1] - The competition in AI video is shifting from generation capability to sustainable creation ability, indicating a deeper industry consensus [1] - The introduction of tools like Seko by SenseTime reflects a move towards integrated content production infrastructure, focusing on multi-episode creation [4] Group 2: Company Developments - SenseTime's Seko platform has attracted over 200,000 creators, including numerous short drama studios, with some works successfully launched on mainstream platforms [5] - Seko supports a combination of self-developed and third-party models, ensuring consistency across different models, which is crucial for building sustainable IP assets [5] - The Seko platform's recent version emphasizes multi-episode creation, addressing the bottleneck of commercializing AI short dramas [4] Group 3: User Experience and Market Dynamics - The competition among platforms is shifting from model parameters to user experience and workflow integration, aiming to simplify the creation process for non-technical users [6] - SenseTime's approach includes a dual-track design that caters to both low-frequency and high-frequency users, allowing for easy content generation while maintaining control for advanced users [6] - The rapid growth of creators on the Seko platform, from 100,000 to over 200,000 in a short period, indicates a shift from consumer to professional use [6] Group 4: Cost Efficiency and Future Outlook - Seko has achieved a 50% reduction in inference costs through model optimization and adaptation to domestic chips, making high-quality video production more accessible [7] - Despite challenges in copyright, content safety, and IP protection, SenseTime aims to enhance creator tools rather than create a closed content platform [7] - The ultimate goal of the technology is to enhance human creativity rather than replace it, positioning Seko as a supportive tool for creators [7]
国产芯片也能跑AI视频实时生成了,商汤Seko 2.0揭秘幕后黑科技
机器之心· 2025-12-15 08:10
Core Insights - The article discusses the competitive landscape of video generation models, highlighting the advancements made by various tech companies, including Google, Runway, and Kuaishou, while questioning the readiness of these models as productivity tools [2][9] - SenseTime's Seko 2.0 version is introduced as a significant advancement, enabling AI short drama creation with minimal human input, effectively allowing a single person to manage the production [2][4][7] Group 1: Industry Developments - Major tech companies are racing to release enhanced versions of video generation models before the end of the year, with Google launching Veo 3.1 and Runway introducing Gen-4.5 [2] - SenseTime's Seko 2.0 has been successfully deployed in over a hundred short drama studios, showcasing its capability to generate scripts, storyboards, and videos rapidly [7][9] Group 2: Technical Challenges - The article outlines the "impossible triangle" of video generation, where efficiency, cost, and quality are at odds, making it difficult for AI video generation models to meet commercial demands [11][13] - Current models, even at the Sora 2 level, require several minutes to generate just 10 seconds of video, which hampers rapid iteration and real-time feedback essential for industrial production [11][12] Group 3: Innovations in Video Generation - SenseTime's LightX2V framework is highlighted as a breakthrough in real-time video generation, achieving generation times of under 5 seconds for 5-second videos, significantly faster than current industry standards [16][17] - The framework employs Phased DMD technology, which enhances video quality and consistency while maintaining high generation speeds [19][20] Group 4: Engineering and Optimization - LightX2V incorporates a comprehensive optimization strategy across five dimensions: model, scheduling, computation, storage, and communication, enabling low-cost and real-time video generation [31][32] - The framework's architecture allows for efficient use of consumer-grade GPUs, achieving real-time generation capabilities with a memory requirement of less than 8GB [36][37] Group 5: Domestic Chip Adaptation - SenseTime's Seko 2.0 has achieved full compatibility with domestic AI chips, allowing for a cost-effective alternative to NVIDIA chips while maintaining comparable video quality [39][40] - The strategic support for domestic AI ecosystems is emphasized, marking a significant step for China's AI industry in achieving core technological independence [42]
迪士尼向OpenAI投资15亿美元 给旗下IP进行AI视频生成
Sou Hu Cai Jing· 2025-12-12 00:44
【CNMO科技消息】当地时间12月11日,据《华尔街日报》报道,华特迪士尼公司已与OpenAI达成一 项总额约15亿美元的战略投资与合作协议。此次合作核心是OpenAI获得迪士尼旗下庞大知识产权的授 权,将其AI视频生成模型Sora向迪士尼的创意内容生态开放。 根据为期三年的独家授权协议,Sora的用户将能够使用超过200个来自迪士尼、漫威、星战、皮克斯等 品牌的经典角色进行短视频创作。部分精选的AI生成内容未来有望在迪士尼流媒体平台Disney+上播 出。 协议规定,OpenAI需为使用迪士尼IP支付授权费用。同时,迪士尼通过此次投资,获得了在未来以当 前约5000亿美元估值增持OpenAI股份的权利。此外,迪士尼员工也将获得使用ChatGPT等OpenAI工具 进行内部产品开发的权限。 值得注意的是,此次合作宣布的时机颇为微妙。就在前一天,迪士尼刚向谷歌发去警告信,指控其AI 图像工具"Gemini"大规模生成并传播涉及迪士尼角色(如达斯·维德、蜘蛛侠等)的侵权内容,要求其 停止侵权。分析认为,迪士尼选择与OpenAI联手,意在对其主要竞争对手谷歌形成制衡。 为了防范IP滥用风险,协议包含了严格的内容安全 ...