Workflow
物理AGI
icon
Search documents
「具身原生」元年!专访原力灵机汪天才,解析具身智能的「PyTorch时刻」
机器之心· 2026-02-10 08:52
Core Viewpoint - The article discusses the significant advancements in embodied intelligence, particularly through the launch of the Dexbotic 2.0 framework and its collaboration with RLinf, marking a pivotal moment in the industry towards a "native embodied" era of AI [3][5][9]. Group 1: Framework and Collaboration - The Dexbotic 2.0 framework aims to standardize the infrastructure for embodied intelligence, similar to how PyTorch revolutionized deep learning [5][16]. - The collaboration with Tsinghua University and RLinf focuses on enhancing the capabilities of embodied AI through a unified framework that integrates perception, decision-making, and execution [3][5][19]. - The introduction of the DM0 model and the DFOL workflow signifies a comprehensive approach to developing and deploying embodied applications [6][51]. Group 2: Embodied Native Concept - "Embodied Native" is defined as a concept that emphasizes a closed-loop system of perception, decision-making, and execution, allowing AI to interact with the physical world effectively [15][13]. - The framework promotes the use of real-world data and multi-modal training to enhance the model's understanding and interaction with its environment [17][41]. - The transition from a "big model brain + mechanical limbs" approach to a fully integrated embodied system is highlighted as a key evolution in the field [12][13]. Group 3: Technical Innovations - Dexbotic 2.0 features a modular design that maintains high flexibility while ensuring end-to-end processing, allowing for independent upgrades of perception, cognition, and control modules [21][33]. - The framework integrates various models and capabilities, including visual-language-action (VLA) and navigation, to achieve comprehensive task execution [37][38]. - The introduction of a standardized data format (Dexdata) and a unified training pipeline addresses the fragmentation in the development of embodied intelligence [45][46]. Group 4: Performance and Evaluation - The DM0 model, with 2.4 billion parameters, has achieved high performance in real-world evaluations, demonstrating its capability in both single and multi-task scenarios [57][58]. - The RoboChallenge benchmark is established to provide a fair evaluation of embodied models, ensuring that performance metrics reflect true capabilities rather than optimized scores [46][57]. - The DFOL workflow enables continuous improvement of robotic systems through real-time data feedback, enhancing their operational efficiency [62][65]. Group 5: Future Insights - The article emphasizes the importance of integrating multi-modal sensory inputs, such as touch and auditory capabilities, to enhance the modeling of the physical world [74]. - The rapid evolution of embodied intelligence is noted, with expectations for significant advancements in the near future, akin to the pace seen in large model developments [73][75]. - The company advocates for an open-source approach to foster collaboration and innovation within the embodied intelligence community, aiming to lower barriers for developers [68][71].
3个月连融5亿!具身智能公司极佳视界完成2亿元A2轮融资,推出物理AGI的原生模型与原生本体
机器人圈· 2025-12-10 09:37
Core Viewpoint - The company Jiga Vision has recently secured 200 million yuan in A2 round financing, marking its fourth round of financing within three months, totaling 500 million yuan in A round series financing [1][3]. Financing and Investment - The latest financing round was led by Dacheng Caizhi, with participation from existing shareholders and several notable investment institutions [1]. - The company has completed multiple financing rounds, including Pre-A, Pre-A+, and A1, demonstrating strong investor confidence and interest in its growth potential [1]. Product Development and Innovation - Jiga Vision focuses on physical AGI (Artificial General Intelligence) and has launched several products, including the GigaWorld platform, GigaBrain, and Maker, which represent a comprehensive suite of physical AI software and hardware [1][6]. - The company has introduced a new generation of physical AGI native body, Maker H01, which is designed for various applications in home, commercial, and light industrial settings [9][12]. Technological Advancements - The company has developed the world's first world model-driven embodied VLA model, GigaBrain-0, and the leading embodied world model, GigaWorld-0, showcasing its technological leadership in the field [3][4]. - GigaBrain-0 enhances 3D spatial perception and structured reasoning capabilities, allowing for more precise navigation and complex task execution [6]. Market Position and Future Outlook - Jiga Vision believes that the next 2-3 years will be a critical window for breakthroughs in physical AGI, with the integration of world models and action models accelerating the arrival of a "ChatGPT moment" in the physical world [2][14]. - The company is actively collaborating with leading clients across various industries, indicating strong market demand and potential for widespread adoption of its technologies [14].
达晨财智领投 极佳视界完成2亿元A2轮融资
Xin Lang Cai Jing· 2025-12-08 15:14
Investment Overview - The company Jijiashijie has recently completed a new round of financing, raising 200 million yuan in Series A2 funding, led by Dacheng Caizhi, with participation from several notable institutions [1][3] - This round of financing follows three previous rounds (Pre-A, Pre-A+, A1) completed within three months, totaling 500 million yuan in Series A funding [1][3] Company Focus and Products - Jijiashijie specializes in general intelligence for the physical world, aiming for physical AGI (Artificial General Intelligence) and has plans to release a corresponding ontology by November 26, 2025 [1][3] - The company's product offerings include the GigaWorld platform (for driving and embodiment), GigaBrain (general embodied brain), and Maker (general embodied ontology), representing a full-stack approach to physical AI [1][3] Model Development - The company has introduced a native paradigm of "world model + action model + reinforcement learning," where each component is driven by the world model [1][3] - The current trend in model architecture is converging towards general action models, with a shift in data sources to real machine data and world model-generated data [2][4] Industry Trends - The company believes that physical AI is entering a new critical era, with the next 2-3 years being a key window for breakthroughs in physical AGI [5] - The advancements in world models and action models are accelerating the arrival of a "ChatGPT moment" in the physical world [5]
极佳视界完成2亿元A2轮融资 达晨、华控领投
Core Insights - The company, 极佳视界, has recently completed a new round of financing amounting to 200 million yuan in its A2 round, led by 达晨财智 and supported by various notable investors [1] - The company has successfully raised a total of 500 million yuan across four financing rounds within three months, indicating strong investor confidence and interest in its business model [1] - Established in 2023, 极佳视界 focuses on physical AI, aiming to develop general intelligence driven by world models, with a product lineup that includes GigaWorld, GigaBrain, and Maker [1] Financing Details - The latest financing round was led by 达晨财智, with participation from existing shareholders and several prominent investment firms [1] - The company has completed a series of financing rounds including Pre-A, Pre-A+, and A1, showcasing a rapid growth trajectory [1] Product and Technology Focus - 极佳视界's product offerings include a world model platform (GigaWorld), a general embodied brain (GigaBrain), and a general embodied ontology (Maker), representing a comprehensive stack of physical AI software and hardware [1] - The company emphasizes a native model paradigm that integrates world models, action models, and reinforcement learning, positioning itself at the forefront of physical AGI development [2] Industry Outlook - The company believes that the next 2-3 years will be a critical window for breakthroughs in physical AGI, with the integration of world models and action models accelerating advancements in the field [2] - The emergence of a scalable closed-loop iteration between sensors, actuators, data collection devices, and general models is seen as increasingly valuable in the context of physical AI [2]
3个月连融5亿!极佳视界A2轮获2亿,推出物理AGI原生模型与本体
3 6 Ke· 2025-12-08 07:56
Core Insights - The company "极佳视界" has recently completed a new round of financing amounting to 200 million yuan, led by 达晨财智, with participation from several notable investors, bringing the total financing to 500 million yuan over four rounds in three months [2][3] - The company focuses on physical AGI (Artificial General Intelligence) and aims to release a corresponding ontology by November 26, 2025, positioning itself for the future of physical AGI [2][3] Financing and Investment - The latest financing round is part of a series of investments that include Pre-A, Pre-A+, and A1 rounds, indicating strong investor confidence and interest in the company's vision and technology [2] - The total amount raised in the A round series is 500 million yuan, showcasing the company's rapid growth and appeal in the investment community [2] Product Development - The company has developed a comprehensive product matrix that includes GigaWorld (world model platform), GigaBrain (general embodied brain), and Maker (general embodied ontology), indicating a systematic approach to the future of physical AI [3] - The original model paradigm proposed by the company integrates "world model + action model + reinforcement learning," with the world model becoming a core driver for data sources, learning methods, and model architecture [3][6] Technological Advancements - The company has released the world's first world model-driven embodied VLA model GigaBrain-0 and the leading embodied world model GigaWorld-0, achieving industry-leading performance [4][5] - GigaBrain-0 enhances 3D spatial perception and structured reasoning capabilities, allowing for more precise navigation and complex operations, outperforming existing state-of-the-art methods [6] Market Position and Future Outlook - The company believes that the next 2-3 years will be a critical window for breakthroughs in physical AGI, with the "ChatGPT moment" for the physical world approaching rapidly [3][16] - The integration of original models and original ontology is seen as key to achieving the company's goals in the physical AGI space [3][16] Team and Expertise - The core team has a decade of experience in physical AI development and has achieved world-class results in both technological innovation and industrial application [4][5] - The founder, Dr. Huang Guan, has a strong background in automation and has led significant projects in the field, contributing to the company's competitive edge [5]
达晨、华控领投,极佳视界A2轮再融2亿,押注“世界模型+行动模型”原生架构
Tai Mei Ti A P P· 2025-12-08 07:17
Group 1 - The company, Jiga Vision, has completed a new round of financing, raising 200 million yuan in Series A2 funding, led by Dashen Caizhi, with participation from several notable investors, bringing the total funding raised in the last three months to 500 million yuan [2] - The founder and CEO, Dr. Huang Guan, has a strong background in AI and robotics, having previously worked at leading research institutions and has been instrumental in the evolution of physical AI from its inception to industrial application [2][3] - Jiga Vision has introduced a new paradigm for artificial general intelligence (AGI) that emphasizes a "world model + action model + reinforcement learning" framework, indicating a shift towards general action models in the industry [3] Group 2 - The company has officially launched two core models for physical AGI: GigaBrain-0, an end-to-end decision control model, and GigaWorld-0, a high-quality world model, along with the Maker H01 robot platform [4] - GigaBrain-0 enhances 3D spatial perception and structured reasoning capabilities, significantly improving navigation accuracy and task execution in complex environments, outperforming current state-of-the-art methods in various benchmarks [5] - GigaWorld-0 generates high-fidelity, controllable, and diverse interactive data, achieving nearly 300% performance improvement in key generalization dimensions, making it a cost-effective solution in the current market [6] Group 3 - Maker H01 is designed for open environments in home, commercial, and light industrial applications, featuring a dual-arm and omnidirectional mobile chassis, capable of performing precise operations and complex tasks [6][7] - The integration of GigaBrain-0, GigaWorld-0, and Maker H01 accelerates the transition of embodied intelligence from the laboratory to scalable applications, marking a significant step towards a reliable and generalizable physical AGI era [7]
智源发布具身数据创新基座,携手行业共筑物理AGI基础设施
具身智能之心· 2025-12-03 03:47
Core Insights - The article discusses the launch of the RoboXstudio platform and the RoboCOIN dataset by the Beijing Zhiyuan Artificial Intelligence Research Institute, aimed at addressing challenges in embodied data production and enhancing research efficiency in embodied intelligence [6][19]. Group 1: Challenges in Embodied Data - Embodied data faces three main challenges: data silos, lack of quality control, and high costs [7][8]. - Data silos arise from non-standardized formats and isolated data tools, complicating data processing [7]. - Quality control issues include frame loss, stuttering, and timestamp misalignment, leading to unreliable data records [8]. - The cost of generating embodied data remains high due to reliance on manual operations and the absence of mature platforms for scalability [8]. Group 2: CoRobot Software Framework - The CoRobot framework was developed to standardize operations, improve quality, and enhance efficiency in embodied data management [10]. - It consists of five components: data collection tools, format conversion tools, data processing tools, data management tools, and model training tools [10]. Group 3: RoboCOIN Dataset - The RoboCOIN dataset is a collaboration involving multiple companies and universities, designed to be the global benchmark for dual-arm robot data [14][16]. - It features the largest number of dual-arm entities, with 180,000 data entries covering over ten scenarios, including industrial and retail applications [16]. - The dataset is noted for its fine-grained labeling and ease of use, facilitated by the CoRobot framework [16]. Group 4: RoboXstudio Platform - The RoboXstudio platform aims to streamline the entire process of data collection, annotation, management, model training, evaluation, and deployment [19][22]. - It supports diverse robot types and tasks, ensuring comprehensive data collection without gaps [22]. - The platform integrates open-source frameworks and multimodal models to reduce operational costs and enhance user accessibility [22]. Group 5: Open Source and Collaboration - The Zhiyuan Institute emphasizes the importance of collaborative innovation in advancing artificial intelligence, with a significant number of downloads of their open-source models [23]. - The RoboCOIN dataset and CoRobot framework are made available to the public to foster industry-wide collaboration and innovation [23][25].
万字长文聊具身智能“成长史”:具身智能跨越了哪些山海,又将奔向哪里
自动驾驶之心· 2025-08-10 03:31
Core Viewpoint - The article discusses the rapid advancements in embodied intelligence and robotics, emphasizing the need for robots to integrate AI with physical capabilities to perform tasks that are currently challenging for them, such as simple actions that children can do [8][9]. Group 1: Evolution of Embodied Intelligence - Over the past decade, embodied intelligence has evolved significantly, with a focus on integrating AI into robots' control systems to enhance their performance in the physical world [9]. - The gap between research prototypes and practical applications is highlighted, with a need for robots to reach a Technology Readiness Level (TRL) of 8 to 9 for industrial acceptance [10]. - Machine learning advancements, including better sensors and algorithms, have led to substantial improvements in robotics, but achieving high success rates in real-world applications remains a challenge [12][14]. Group 2: Opportunities and Challenges in Robotics - The current landscape presents both opportunities and challenges for robotics, with a focus on structured environments for initial applications before tackling more complex, unstructured settings [14][17]. - The importance of scalable learning systems in robotics is emphasized, as researchers aim to leverage data from multiple robots to enhance performance across various tasks [20]. Group 3: Specialized vs. General Intelligence - The discussion contrasts Artificial Specialized Intelligence (ASI) with Artificial General Intelligence (AGI), suggesting that while ASI focuses on high performance in specific tasks, AGI aims for broader capabilities [27][29]. - The advantages of specialized models include efficiency, robustness, and the ability to run on-premise, while general models offer greater flexibility but are more complex and costly to operate [31][35]. Group 4: Future Directions in Robotics - The emergence of visual-language-action (VLA) models, such as RT-2, represents a significant step forward in robotics, allowing for more complex task execution through remote API calls [44][46]. - The development of the RTX dataset, which includes diverse robotic data, has shown that cross-embodied models can outperform specialized models in various tasks, indicating the potential for generalization in robotics [47][48]. - The second-generation VLA models, like PI-Zero, are designed to handle continuous actions and complex tasks, showcasing advancements in robot dexterity and adaptability [49][50]. Group 5: Data and Performance in Robotics - The importance of data in achieving high performance in robotics is underscored, with a call for large-scale data collection to support the development of robust robotic systems [62][70]. - The article concludes with a discussion on the need for a balance between performance and generalization in robotics, suggesting that achieving high performance is crucial for real-world deployment [66][68].
对话智源王仲远:机器人的大小脑可能会“合体”,但不是今天
AI前线· 2025-06-11 08:39
Core Insights - The article discusses the launch of the "Wujie" series of large models by Zhiyuan Research Institute, focusing on advancements in multi-modal AI technology and its applications in physical AGI [1][2][3] Group 1: New Model Launch - The "Wujie" series includes several models such as Emu3, Brainμ, RoboOS2.0, RoboBrain2.0, and OpenComplex2, aimed at enhancing AI's understanding and interaction with the physical world [1][2] - Emu3 is designed as a native multi-modal architecture that enables large models to comprehend and reason about the world, set to be released in October 2024 [3][4] Group 2: Technological Advancements - Brainμ, based on Emu3, integrates various brain signals to perform multiple neuroscience tasks, demonstrating significant performance improvements over existing models [4][5] - RoboOS2.0 is the first open-source framework for embodied intelligence, allowing seamless integration of skills from various robot models, with a 30% performance enhancement compared to its predecessor [6][7] Group 3: Applications and Collaborations - Brainμ has potential applications in brain-computer interfaces, having successfully reconstructed sensory signals using portable EEG systems [5] - The OpenComplex2 model represents a breakthrough in dynamic conformational modeling of biological molecules, enhancing the understanding of molecular interactions at atomic resolution [11][12] Group 4: Future Directions - The article emphasizes the ongoing evolution of large model technology, with a focus on bridging the gap between digital and physical worlds, which is crucial for achieving physical AGI [2][3] - RoboBrain2.0 has improved task planning and spatial reasoning capabilities, achieving a 74% increase in task planning accuracy compared to its predecessor [8][9]
对话智源研究院院长王仲远:AI正加速从数字世界走向物理世界
Core Insights - The rapid advancement of AI technology is shifting from digital to physical applications, with a focus on humanoid robots as practical tools rather than mere mascots [1][2] - The development trajectory of large models is moving towards multi-modal world models, which aim to enhance AI's understanding and interaction with the physical world [2][3] AI Technology Development - The performance of large language models is reaching a bottleneck, necessitating improvements through reinforcement learning, high-quality synthetic data, and activation of underutilized multi-modal data [1][2] - The introduction of the "Wujie" series of large models, including the Emu3 multi-modal world model, signifies a strategic shift towards understanding physical causal relationships [2][3] Embodied Intelligence - Humanoid robots are recognized for their long-term value due to their design compatibility with human environments and the availability of extensive human behavior data for model training [3][4] - The current limitations in data volume hinder the training of models that integrate both "big brain" and "small brain" functionalities, indicating a need for further development [4][6] Industry Trends - The focus on embodied intelligence is expected to prioritize applications in controlled environments, such as logistics and repetitive tasks, where safety and efficiency are paramount [3][4] - The concept of "big brain" and "small brain" integration is acknowledged as a potential future trend, but current data limitations prevent immediate implementation [4][5] AGI Development - The emergence of Agents in AI signifies a new phase where foundational models can support the development of various applications, akin to mobile apps in the internet era [5][6] - The industry is still in the early stages of embodied intelligence development, facing challenges similar to those encountered in the early days of AI large models [5][6]