Workflow
大型行为模型(LBM)
icon
Search documents
波士顿动力机器人终于有脑子了!人类故意使绊子也不怕
量子位· 2025-08-22 02:30
Core Viewpoint - Boston Dynamics has upgraded its Atlas robot to incorporate end-to-end AI capabilities, allowing it to understand natural language commands, autonomously plan actions, and handle unexpected situations [1][8]. Group 1: Atlas Robot Capabilities - The new version, Atlas MTS, can recognize and open boxes even when the lid is closed [2]. - It can accurately identify changes in the position of objects, such as boxes being moved [4]. - Atlas can discover and correctly place missing items into boxes, showcasing its advanced perception [6]. - The robot can autonomously respond to unexpected situations, such as parts falling or lids not being closed [21][22]. - It has the ability to learn any action that a human can demonstrate, including tying knots and folding chairs [23]. Group 2: Technical Innovations - The upgrade was developed in collaboration with Toyota Research Institute and is based on a Large Behavior Model (LBM) [8]. - The LBM utilizes a diffusion Transformer model with 450 million parameters, converting various inputs like images and natural language into action commands for Atlas [17]. - The integration of a model predictive controller with a VR interface allows for precise control over a wide range of tasks, from fine motor skills to full-body movements [19][20]. Group 3: Transition to Electric Drive - Boston Dynamics has retired the hydraulic version of Atlas and released an all-electric version, which is more cost-effective and integrates better with AI systems [28][29]. - Electric drive systems offer higher precision, lower energy consumption, and are more compatible with AI learning frameworks [30]. - The transition to electric drive has enabled Boston Dynamics to continuously introduce new movements and capabilities for the robot [31][36]. Group 4: Competitive Landscape - The article also mentions the domestic company Yushu, which has consistently used electric drive technology in its robots, achieving rapid iterations and gaining global recognition [39]. - Yushu's product lineup includes various humanoid robots with different specifications and price points, showcasing a focus on electric drive philosophy [41]. Group 5: Future Outlook - The integration of electric drive technology with AI algorithms is expected to usher in a new era for electric robots [44].
机器人「GPT时刻」来了?丰田研究院悄悄做了一场最严谨的VLA验证
具身智能之心· 2025-07-21 08:42
Core Viewpoint - The article discusses the advancements in robotic arms, particularly focusing on the development of Large Behavior Models (LBM) that enable robots to perform complex tasks autonomously, showcasing significant improvements in performance and capabilities compared to traditional models [3][7][15]. Summary by Sections Introduction to Robotic Arms - Robotic arms are typically associated with simple tasks like grabbing or serving ice cream, but the complexity increases exponentially when tasked with more intricate operations such as setting a table or assembling a bicycle [2][3]. Development of VLA Models - The recent progress in Visual-Language-Action (VLA) models has allowed robots to integrate multimodal information (images, instructions, scene semantics) and execute complex tasks, moving towards more intelligent and versatile systems [3][4]. Large Behavior Models (LBM) - LBM represents a significant advancement in robotic capabilities, built on diffusion model strategies, enabling robots to autonomously execute complex operations with impressive results [7][10][19]. - The research conducted by Toyota Research Institute (TRI) and led by notable scholars emphasizes the rigorous evaluation of these models, demonstrating their effectiveness in both simulated and real-world environments [9][10]. Training and Evaluation - The LBM was trained on a diverse dataset, including 1,700 hours of robot data, and underwent 1,800 real-world evaluations and over 47,000 simulated deployments, showcasing its robust performance [13][14]. - The findings indicate that even with limited training data, the model's performance significantly improves, suggesting a positive trend towards achieving effective data acquisition and performance enhancement [14][16]. Performance Metrics - The evaluation metrics included success rate and task completion, with a focus on relative success rates to better compare different methods' performances [26][27]. - The LBM demonstrated superior performance in both seen and unseen tasks compared to single-task baseline models, indicating its robustness and adaptability [31][39]. Conclusion and Future Implications - The research suggests that the advent of general large-scale models in robotics is on the horizon, hinting at a potential "GPT moment" for embodied intelligence [15][43]. - The results indicate that pre-training can lead to better task performance with less data, reinforcing the idea that as data volume increases, performance benefits will continue to manifest [43][45].