Workflow
RynnRCP
icon
Search documents
达摩院开源具身大脑基模RynnBrain,登顶16项榜单,超越Gemini
Jin Rong Jie· 2026-02-10 02:56
Core Insights - Alibaba's Damo Academy has released the RynnBrain model, which significantly enhances robotic intelligence by introducing spatiotemporal memory and spatial reasoning capabilities, surpassing leading models like Google's Gemini Robotics ER 1.5 in 16 benchmarks [1][4][6] Model Features - RynnBrain incorporates spatiotemporal memory, allowing robots to locate objects and predict movement trajectories based on their historical memory, thus providing global temporal recall capabilities [2][7] - The model employs a novel reasoning strategy that intertwines text and spatial positioning, reducing hallucination issues and ensuring that reasoning is grounded in the physical environment [2][7] Performance Metrics - RynnBrain has achieved state-of-the-art (SOTA) results across 16 embodied evaluation benchmarks, including environmental perception, object reasoning, and trajectory prediction, outperforming models from Google and NVIDIA [4][6] - The model was trained on over 20 million pairs of data using the proprietary RynnScale architecture, which accelerates training by two times with the same resources [4][7] Scalability and Applications - RynnBrain demonstrates excellent scalability, enabling rapid fine-tuning for various embodied models such as navigation and planning with minimal data, achieving superior performance compared to Gemini 3 Pro [7] - The model is open-sourced, including a complete set of reasoning and training codes, and introduces the RynnBrain-Bench for evaluating fine-grained spatiotemporal tasks, filling a gap in the industry [7][9] Industry Impact - The introduction of RynnBrain marks a significant advancement in embodied intelligence, facilitating the transition of AI from digital environments to real-world applications [9]
机器人上下文协议首次开源:阿里达摩院一口气放出具身智能「三大件」
具身智能之心· 2025-08-12 00:03
Core Viewpoint - Alibaba's Damo Academy announced the open-source of several models and protocols aimed at enhancing embodied intelligence, addressing challenges in data, model, and robot compatibility, and streamlining the development process [1][2]. Group 1: Open-Source Models and Protocols - The RynnRCP protocol was introduced to facilitate the integration of various data, models, and robotic systems, creating a seamless workflow from data collection to action execution [2][5]. - RynnVLA-001 is a visual-language-action model that learns human operational skills from first-person perspective videos, enabling smoother robotic arm control [7]. - The RynnEC model incorporates multi-modal large language capabilities, allowing for comprehensive scene analysis across 11 dimensions, enhancing object recognition and interaction in complex environments [7]. Group 2: Technical Framework and Features - The RCP framework connects robotic bodies with sensors, providing standardized interfaces and compatibility across different transport layers and model services [5]. - RobotMotion serves as a bridge between large models and robotic control, converting low-frequency commands into high-frequency control signals for smoother robot movements [5][6]. - The framework includes integrated simulation and real-machine control tools, facilitating quick developer onboarding and supporting various functionalities like task regulation and trajectory visualization [5]. Group 3: Industry Engagement and Community Building - Damo Academy is actively investing in embodied intelligence, focusing on system and model development, and collaborating with various stakeholders to build industry infrastructure [7]. - The launch of the WorldVLA model, which merges world models with action models, has garnered significant attention for its enhanced understanding and generation capabilities [8]. - The establishment of the "Embodied Intelligence Heart" community aims to foster collaboration among developers and researchers in the field, providing resources and support for various technical directions [11][12].
腾讯研究院AI速递 20250812
腾讯研究院· 2025-08-11 16:01
Group 1 - xAI announced the free global availability of Grok 4, limiting usage to 5 times every 12 hours, which has led to dissatisfaction among paid users who feel betrayed by the subscription model [1] - Inspur released the "Yuan Nao SD200" super-node AI server, integrating 64 cards into a unified memory system, capable of running multiple domestic open-source models simultaneously [2] - Zhiyuan published the GLM-4.5 technical report, revealing details on pre-training and post-training, achieving native integration of reasoning, coding, and agent capabilities in a single model [3] Group 2 - Kunlun Wanwei launched the SkyReels-A3 model, capable of generating high-quality digital human videos up to one minute long, optimized for hand motion interaction and camera control [4] - Chuangxiang Sanwei partnered with Tencent Cloud to enhance 3D generation capabilities for its AI modeling platform MakeNow, utilizing Tencent's mixed model [5][6] - Alibaba's DAMO Academy open-sourced three core components for embodied intelligence, including a visual-language-action model and a robot context protocol [7] Group 3 - Baichuan Intelligent released the 32B parameter medical enhancement model Baichuan-M2, outperforming all open-source models in the OpenAI HealthBench evaluation, second only to GPT-5 [8] - Lingqiao Intelligent showcased the DexHand021 Pro, a highly dexterous robotic hand with 22 degrees of freedom, designed to simulate human hand functions accurately [9] - A report indicated that 45% of enterprises have deployed large models in production, with users averaging 4.7 different products, highlighting low brand loyalty in a competitive landscape [10][12]
达摩院开源具身智能“三大件” 机器人上下文协议首次开源
Huan Qiu Wang· 2025-08-11 04:17
Core Insights - Alibaba's Damo Academy announced the open-source release of several models and protocols aimed at enhancing the compatibility and adaptability of data, models, and robots in the field of embodied intelligence [1][3] - The introduction of the Robotics Context Protocol (RynnRCP) aims to address challenges such as fragmented development processes and difficulties in adapting data and models to robotic systems [1][2] Group 1: Open-source Models and Protocols - The RynnVLA-001 model is a visual-language-action model that learns human operational skills from first-person perspective videos, enabling smoother robotic arm control [3] - The RynnEC model integrates multi-modal large language capabilities, allowing for comprehensive scene analysis across 11 dimensions, enhancing object localization and segmentation in complex environments [3] - RynnRCP serves as a complete robot service protocol and framework, facilitating the workflow from sensor data collection to model inference and robotic action execution [1][2] Group 2: Technical Framework and Features - The RCP framework within RynnRCP establishes connections between robotic bodies and sensors, providing standardized capability interfaces and compatibility across different transport layers and model services [2] - The RobotMotion module acts as a bridge between large models and robotic control, converting low-frequency inference commands into high-frequency continuous control signals for smoother robotic movements [2] - The integrated simulation-real machine control tool within RobotMotion aids developers in quickly adapting to tasks, supporting simulation synchronization, data collection, playback, and trajectory visualization [2] Group 3: Industry Engagement and Development - Damo Academy is actively investing in embodied intelligence, focusing on system and model development while collaborating with various stakeholders to build industry infrastructure [3] - The recent open-sourcing of the WorldVLA model, which merges world models with action models, has garnered significant attention for its enhanced understanding and generation capabilities in images and actions [3]
机器人上下文协议首次开源:阿里达摩院一口气放出具身智能「三大件」
机器之心· 2025-08-11 03:19
Core Viewpoint - Alibaba's Damo Academy announced the open-source release of several models and protocols aimed at enhancing the compatibility and integration of data, models, and robots in the field of embodied intelligence, addressing significant challenges in the development process [2][8]. Group 1: Open-source Models and Protocols - The open-source models include RynnVLA-001-7B, a visual-language-action model, RynnEC, a world understanding model, and RynnRCP, a robotic context protocol [2][8]. - RynnRCP aims to facilitate the connection between robot bodies and sensors, providing standardized capability interfaces and ensuring compatibility between different transport layers and model services [6][8]. - RynnVLA-001 is designed to learn human operational skills from first-person perspective videos, enabling smoother and more coherent control of robotic arms [8]. Group 2: RynnRCP Framework and RobotMotion - RynnRCP consists of two main modules: the RCP framework and RobotMotion [5]. - RobotMotion serves as a bridge between large embodied models and robot body control, converting discrete low-frequency inference commands into continuous high-frequency control signals for smooth robotic movement [7]. - The framework supports various popular models and robotic arms, including Pi0, GR00T N1.5, SO-100, and SO-101, and is continuously expanding its capabilities [2][8]. Group 3: Industry Development and Collaboration - Damo Academy is actively investing in embodied intelligence, focusing on system and model development, and collaborating with various parties to build industry infrastructure, including hardware compatibility and data collection [8]. - The recent open-source release of the WorldVLA model, which integrates world models with action models, has garnered significant attention in the industry for its enhanced understanding and generation capabilities [9].