宇树Go1机器人
Search documents
仅看视频就能copy人类动作,宇树G1分分钟掌握100+,UC伯克利提出机器人训练新方式
量子位· 2025-05-08 04:04
Core Viewpoint - The article discusses the development of a new robotic training system called VideoMimic by a team from UC Berkeley, which allows robots to learn human movements from video without the need for motion capture technology [1][2]. Group 1: VideoMimic System Overview - VideoMimic has successfully enabled the Yushun G1 robot to mimic over 100 human actions [2]. - The core principle of VideoMimic involves extracting pose and point cloud data from videos, training in a simulated environment, and ultimately transferring the learned actions to a physical robot [3][17]. - The system has garnered significant attention online, with comparisons made to characters like Jack Sparrow from "Pirates of the Caribbean" [4]. Group 2: Training Process - The research team collected a dataset of 123 video clips filmed in everyday environments, showcasing various human movement skills and scenarios [5][6]. - The Yushun Go1 robot has been trained to adapt to different terrains and perform actions such as stepping over curbs and descending stairs, demonstrating its ability to maintain balance even when slipping [7][14][16]. Group 3: Technical Workflow - VideoMimic's workflow consists of three main steps: converting video to a simulation environment, training control strategies in simulation, and validating these strategies on real robots [18]. - The first step involves reconstructing human motion and scene geometry from single RGB videos, optimizing for accurate alignment of human movements and scene geometry [19]. - The second step processes the scene point cloud into a lightweight triangular mesh model for efficient collision detection and rendering [21]. Group 4: Strategy Training and Deployment - The training process is divided into four progressive stages, resulting in a robust control strategy that requires only the robot's proprioceptive information and local height maps as input [24]. - The Yushun Go1 robot, equipped with 12 degrees of freedom and various sensors, serves as the physical testing platform for deploying the trained strategies [30][31]. - The deployment involves configuring the robot's PD controller to match the simulation environment and utilizing real-time data from its depth camera and IMU for effective movement [35][39]. Group 5: Research Team - The project features four co-authors, all PhD students at UC Berkeley, with diverse research interests in robotics, computer vision, and machine learning [43][48][52].