让机器人拥有“本能”,北京人形开源VLA模型XR-1全链条生态

Core Insights - Beijing Humanoid Robot Innovation Center has launched the XR-1 model, the first domestic VLA model to pass the embodied intelligence national standard test, along with RoboMIND 2.0 and ArtVIP for data training support [1] Group 1: XR-1 Model Capabilities - XR-1 features three core capabilities: cross-data source learning, cross-modal alignment, and cross-ontology control [1] - Cross-data source learning allows the robot to train using vast amounts of human video, reducing training costs and improving efficiency [1] - Cross-modal alignment enables the robot to integrate visual perception and actions, achieving true knowledge-action unity [1] - Cross-ontology control allows XR-1 to quickly match different types and brands of robotic bodies [1] Group 2: Technology and Training - The UVMC (Unified Representation of Multi-modal Vision and Motion) technology developed by Beijing Humanoid creates a mapping bridge between vision and action, enabling robots to respond instinctively to visual stimuli [1] - The multi-configuration pre-training capability of TianGong 2.0 allows for full-body multi-joint humanoid control, enabling significant bending and precise grasping of randomly placed material boxes [1]