Workflow
昆仑万维正式发布SkyReels-A3模型

Core Insights - Kunlun Wanwei officially launched the SkyReels-A3 model on August 11, which utilizes a combination of Diffusion Transformer video diffusion model, frame interpolation model, reinforcement learning for action optimization, and controllable camera movement to create audio-driven digital humans for personalized and interactive content creation [1][2] - The SkyWork AI technology release week commenced on August 11, with Kunlun Wanwei set to unveil a new model each day for five consecutive days, covering various cutting-edge multi-modal AI applications [1] Group 1 - The SkyReels-A3 model allows users to upload a portrait image and a voice segment, enabling the digital human to speak or sing according to the audio input [2] - The model can also replace the original video's audio, automatically synchronizing the character's lip movements, expressions, and performance with the new audio while maintaining visual continuity [2] - Kunlun Wanwei has optimized the model for specific applications such as online live streaming, focusing on the naturalness and clarity of interactive actions in longer, consistent videos [2] Group 2 - For artistic applications like music videos, movie clips, or speeches, Kunlun Wanwei developed a camera control module based on ControlNet structure, allowing for precise frame-level camera movement control [2] - This camera control module extracts depth information from reference images and uses camera parameters to render target camera movement trajectories, guiding the model to reproduce accurate camera effects in generated digital human videos [2] - The performance of SkyReels-A3 has been validated through extensive experiments, showcasing its capabilities in audio-driven video generation compared to existing state-of-the-art models [3]