Diffusion Policy

Search documents
NIPS 2025 MARS 多智能体具身智能挑战赛正式启动!
具身智能之心· 2025-08-18 00:07
Core Insights - The article discusses the challenges and advancements in multi-agent embodied intelligence, emphasizing the need for efficient collaboration among robotic systems to tackle complex tasks in real-world environments [3][4]. Group 1: Challenges in Embodied Intelligence - Single intelligent agents are insufficient for complex and dynamic task scenarios, necessitating high-level collaboration among multiple embodied agents [3]. - The MARS Challenge aims to address these challenges by encouraging global researchers to explore high-level planning and low-level control capabilities of multi-agent systems [4]. Group 2: MARS Challenge Overview - The MARS Challenge features two complementary tracks focusing on planning and control, aiming to evaluate the capabilities of intelligent agents in complex tasks [4][12]. - The challenge will culminate in results and awards announced at the NeurIPS 2025 SpaVLE Workshop [4]. Group 3: Track 1 - Multi-Agent Embodied Planning - Track 1 focuses on high-level task planning and role assignment for heterogeneous robots, utilizing the ManiSkill platform and RoboCasa dataset [5][6]. - Participants will use visual language models to select appropriate robot combinations and create high-level action sequences based on natural language instructions [5][8]. Group 4: Track 2 - Multi-Agent Control Strategy Execution - Track 2 emphasizes the collaborative capabilities of multi-agent systems in executing complex tasks, requiring real-time interaction with dynamic environments [12]. - The RoboFactory simulation environment will be used to develop and evaluate cooperative strategies, with participants designing deployable control models [12][13]. Group 5: Timeline and Participation - The challenge timeline includes a warm-up round starting on August 18, 2025, and the official competition beginning on September 1, 2025, concluding on October 31, 2025 [25]. - Participants from various fields such as robotics, computer vision, and natural language processing are encouraged to join and showcase their creativity and technology [26].
VLA之外,具身+VA工作汇总
具身智能之心· 2025-07-14 02:21
Core Insights - The article focuses on advancements in embodied intelligence and robotic manipulation, highlighting various research projects and methodologies aimed at improving robotic capabilities in real-world applications [2][3][4]. Group 1: 2025 Research Initiatives - Numerous projects are outlined for 2025, including "Steering Your Diffusion Policy with Latent Space Reinforcement Learning" and "Chain-of-Action: Trajectory Autoregressive Modeling for Robotic Manipulation," which aim to enhance robotic manipulation through advanced learning techniques [2][3]. - The "BEHAVIOR Robot Suite" is designed to streamline real-world whole-body manipulation for everyday household activities, indicating a focus on practical applications of robotics [2]. - "You Only Teach Once: Learn One-Shot Bimanual Robotic Manipulation from Video Demonstrations" emphasizes the potential for efficient learning methods in robotic training [2][3]. Group 2: Methodologies and Techniques - The article discusses various methodologies such as "Adaptive 3D Scene Representation for Domain Transfer in Imitation Learning" and "Learning the RoPEs: Better 2D and 3D Position Encodings with STRING," which aim to improve the adaptability and efficiency of robotic systems [2][3][4]. - "RoboGrasp: A Universal Grasping Policy for Robust Robotic Control" highlights the development of a versatile grasping policy that can be applied across different robotic platforms [2][3]. - "Learning Dexterous In-Hand Manipulation with Multifingered Hands via Visuomotor Diffusion" showcases advancements in fine motor skills for robots, crucial for complex tasks [4]. Group 3: Future Directions - The research emphasizes the importance of integrating visual and tactile feedback in robotic systems, as seen in projects like "Adaptive Visuo-Tactile Fusion with Predictive Force Attention for Dexterous Manipulation" [7]. - "Zero-Shot Visual Generalization in Robot Manipulation" indicates a trend towards developing robots that can generalize learned skills to new, unseen scenarios without additional training [7]. - The focus on "Human-to-Robot Data Augmentation for Robot Pre-training from Videos" suggests a shift towards leveraging human demonstrations to enhance robotic learning processes [7].