突破2D-3D鸿沟!北大提出VIPA-VLA,视频解锁机器人精准操控
具身智能之心·2025-12-26 00:55

Core Insights - The article discusses a new approach to robot learning that addresses the challenge of aligning 2D visual information with 3D spatial understanding, which has been a significant limitation in existing visual-language-action (VLA) models [3][6][41] - The research introduces a novel pre-training paradigm that utilizes human demonstration videos to enhance robots' spatial perception capabilities, allowing them to infer 3D spatial relationships from 2D visual inputs [4][40] Research Background - Current VLA models face limitations due to reliance on expensive robot datasets and lack of explicit 3D spatial modeling, which hampers their ability to accurately map physical actions [6][7] - Human demonstration videos provide a solution by offering diverse scenarios and inherent visual-physical correspondences that serve as valuable supervision signals for robot learning [7][8] Hand3D Dataset - The Hand3D dataset, comprising Hand3D-visual and Hand3D-action components, is described as a "3D spatial textbook" for robots, enabling them to learn visual-physical alignment [8][9] - The dataset includes data from nine heterogeneous human manipulation datasets, ensuring a wide variety of scenes and tasks [8][9] Model Architecture: VIPA-VLA - The VIPA-VLA model features a dual-encoder architecture that integrates semantic visual features with 3D spatial features, enhancing the model's ability to understand both scene semantics and spatial structures [15][20] - The model employs a cross-attention fusion layer to combine these features, allowing for effective learning of 3D relationships from 2D inputs [17][20] Training Process - The training process consists of three phases: 3D visual pre-training, 3D action pre-training, and post-training for task adaptation, ensuring a gradual acquisition of 3D capabilities [21][22] - The first phase focuses on aligning semantic and spatial features, while the second phase teaches the model to predict 3D motion tokens based on visual-language inputs [22][23] Experimental Results - VIPA-VLA outperformed existing baselines in various tasks, achieving a success rate of 92.4% in single-view settings and 96.8% in dual-view settings on the LIBERO benchmark [27][28] - In the RoboCasa benchmark, VIPA-VLA achieved a success rate of 45.8%, surpassing other models, particularly in tasks requiring precise 3D positioning [30] - The model demonstrated strong performance in real-world tasks, achieving a 60% success rate in the Wipe-Board task, significantly higher than competing models [31][34] Significance and Future Directions - The research presents a new paradigm for robot learning that reduces reliance on costly robot data and enhances model generalization by leveraging human demonstration videos [40][41] - Future work aims to combine this pre-training paradigm with robot data pre-training and expand the Hand3D dataset to include more complex human-robot interaction tasks [40][41]

突破2D-3D鸿沟!北大提出VIPA-VLA,视频解锁机器人精准操控 - Reportify