夸克AI实验室与浙大联合开源OmniAvatar:音频驱动全身视频生成新突破
Guan Cha Zhe Wang·2025-07-25 04:16

Core Insights - Quark AI Technology Team has partnered with Zhejiang University to open-source OmniAvatar, an innovative audio-driven full-body video generation model that promises revolutionary changes in the video generation field [1] Group 1: Technology Advancements - OmniAvatar overcomes traditional limitations by enabling full-body motion driven by audio, rather than just facial movements, allowing for precise control [1] - The model generates videos by inputting a single image and an audio clip, significantly enhancing lip-sync details and the fluidity of full-body movements [1] - OmniAvatar incorporates a pixel-based audio embedding strategy, allowing audio features to be integrated at a pixel level within the model's latent space, resulting in more natural body movements [2] Group 2: Challenges and Solutions - Long video generation has been a challenge in audio-driven video creation; OmniAvatar addresses this with image embedding strategies and frame overlap techniques to ensure video coherence and consistent character identity [1] - A balance fine-tuning strategy based on LoRA has been proposed to efficiently adapt the model without altering its underlying capacity, allowing it to learn audio features while maintaining video quality and detail [2] Group 3: Future Directions - OmniAvatar represents an initial attempt in multi-modal video generation, having shown preliminary validation on experimental datasets but not yet reaching product-level application [2] - Future explorations will focus on enhancing complex instruction processing capabilities and multi-character interactions to broaden the model's applicability in various scenarios [2]

夸克AI实验室与浙大联合开源OmniAvatar:音频驱动全身视频生成新突破 - Reportify