Workflow
具身智能之心
icon
Search documents
让机器人「不只是走路」,Nav-R1引领带推理的导航新时代
具身智能之心· 2025-09-19 00:03
Core Viewpoint - The article discusses the introduction of Nav-R1, a new embodied foundation model designed to enhance the reasoning and navigation capabilities of robots in 3D environments, integrating perception, reasoning, and action effectively [5][30]. Group 1: Key Innovations - Nav-R1 utilizes a large-scale dataset called Nav-CoT-110K, which contains approximately 110,000 Chain-of-Thought trajectories, facilitating a stable reasoning and action foundation before reinforcement learning optimization [8][6]. - The model incorporates three types of rewards: Format Reward, Understanding Reward, and Navigation Reward, which ensure structured output, semantic understanding, and path fidelity respectively [10][15]. - The Fast-in-Slow reasoning paradigm is inspired by human cognition, where a fast system handles immediate responses while a slow system manages long-term planning and semantic consistency [11][16]. Group 2: Experimental Results - Nav-R1 demonstrated significant improvements in various navigation tasks, achieving an increase of approximately 8% or more in success rates and path efficiency compared to other advanced methods [14]. - In real-world deployments, Nav-R1 was tested on a mobile robot platform, showing robust performance in navigating complex indoor environments [19][26]. Group 3: Applications and Implications - The model has potential applications in service robots and home assistants, enhancing user experience by enabling robots to navigate cluttered environments and understand commands [31]. - In healthcare settings, Nav-R1 can assist in navigating complex environments safely and reliably, which is crucial for elderly care and medical facilities [32]. - The technology is also applicable in augmented and virtual reality, where virtual agents need to navigate physical spaces effectively [33]. - In industrial and hazardous environments, Nav-R1's robustness and generalization capabilities make it suitable for executing tasks in unknown or dangerous settings [34].
具身的这几个方向,组成了所谓的大小脑算法
具身智能之心· 2025-09-19 00:03
Core Viewpoint - The article discusses the evolution and current trends in embodied intelligence technology, emphasizing the integration of various models and techniques to enhance robotic capabilities in real-world environments [3][10]. Group 1: Technology Development Stages - The development of embodied intelligence has progressed through several stages, starting from grasp pose detection to behavior cloning, and now to diffusion policy and VLA models [7][10]. - The first stage focused on static object grasping with limited decision-making capabilities [7]. - The second stage introduced behavior cloning, allowing robots to learn from expert demonstrations but faced challenges in generalization and error accumulation [7]. - The third stage, marked by the introduction of diffusion policy methods, improved stability and generalization by modeling action sequences [8]. - The fourth stage, beginning in 2025, explores the integration of VLA models with reinforcement learning and world models to enhance predictive capabilities and multi-modal perception [9][10]. Group 2: Key Technologies and Techniques - Key technologies in embodied intelligence include VLA, diffusion policy, and reinforcement learning, which collectively enhance robots' task execution and adaptability [5][10]. - VLA models combine visual perception, language understanding, and action generation, enabling robots to interpret human commands and perform complex tasks [8]. - The integration of tactile sensing with VLA models expands the sensory capabilities of robots, allowing for more precise operations in unstructured environments [10]. Group 3: Industry Implications and Opportunities - The advancements in embodied intelligence are leading to increased demand for engineering and system capabilities, transitioning from theoretical research to practical deployment [10][14]. - There is a growing interest in training and deploying various models, including diffusion policy and VLA, on platforms like Mujoco and IsaacGym [14]. - The industry is witnessing a surge in job opportunities and research interest, prompting many professionals to shift focus towards embodied intelligence [10].
390亿美元,全球具身智能第一估值来了!英伟达持续加注中
具身智能之心· 2025-09-19 00:03
Core Insights - Figure has successfully raised over $1 billion in Series C funding, achieving a post-money valuation of $39 billion, setting a record in the field of embodied intelligence [3][33] - The funding round was led by Parkway Venture Capital, with participation from major investors including Nvidia, Brookfield Asset Management, and Intel Capital [5] - The company aims to expand its humanoid robot manufacturing and deployment in both household and commercial settings [10][22] Funding and Valuation - The Series C funding raised over $1 billion, resulting in a valuation of $39 billion, the highest in the current publicly available information on the embodied intelligence sector [3][33] - Previous funding rounds include a $675 million Series B round in February 2024, which valued the company at $2.6 billion [23] Technological Advancements - Figure has developed the Helix architecture, a visual-language-action model that allows robots to perceive, understand, and act like humans [18][22] - The Helix system consists of two components that communicate with each other, enabling the robot to perform various tasks using a unified model [19] - The latest funding will support the development of the Helix system, including building next-generation GPU infrastructure and advanced data collection projects [22][21] Recruitment and Expansion - Figure is actively recruiting across 13 areas, including AI-Helix and BotQ manufacturing, to support its growth and technological advancements [6] - The company is expanding its humanoid robot production capabilities to assist with household chores and commercial labor tasks [10][22] Market Position - Figure has positioned itself as a leading player in the humanoid robotics sector, especially after parting ways with OpenAI and focusing on developing its proprietary AI models [29][31] - The company has quickly gained attention in the market, with significant advancements in technology and funding, making it a notable competitor in the embodied intelligence landscape [32][33]
VLA的论文占据具身方向的近一半......
具身智能之心· 2025-09-18 04:00
Core Insights - The article emphasizes the significance of Vision-Language-Action (VLA) models in the field of embodied intelligence, highlighting their ability to enable robots to autonomously make decisions in diverse environments, thus breaking the limitations of traditional single-task training methods [1][4]. Industry Development - The embodied intelligence sector is experiencing rapid growth, with teams like Unitree, Zhiyuan, Xinghaitu, and Yinhai General transitioning from laboratory research to commercialization, alongside major tech companies such as Huawei, JD, and Tencent collaborating with international firms like Tesla and Figure AI [3]. Research Opportunities - VLA is identified as a current research hotspot with many unresolved issues, making it a promising area for academic papers. The article mentions the establishment of a specialized VLA research guidance course aimed at helping individuals quickly enter or transition within this field [3][4]. Course Content and Structure - The course focuses on how agents interact effectively with the physical world through a perception-cognition-action loop, covering the evolution of VLA technology from early grasp pose detection to recent models like Diffusion Policy and multimodal foundational models [7][8]. - It addresses core challenges in embodied intelligence, such as cross-domain generalization and long-term planning, and explores how to integrate large language models with robotic control systems [8]. Learning Outcomes - Upon completion, participants are expected to master the theoretical foundations and technical evolution of VLA models, gain proficiency in simulation environments, and develop independent research capabilities [14]. - The course aims to guide students from idea generation to the completion of a high-quality academic paper, ensuring they can identify research opportunities and design effective experiments [10][14].
10000台,特斯拉Optimus Gen3刚刚拿下了全球最大订单!
具身智能之心· 2025-09-18 01:23
Core Insights - Tesla's Optimus Gen3 has secured its first external order for 10,000 units from PharmAGRI, aimed at automating drug production processes for precision and efficiency [1] - Elon Musk invested $10 billion in Tesla stock, linked to a performance-based compensation plan that could unlock $1.2 trillion in stock rewards if 1 million units of Optimus are delivered in the future [1] - The Optimus Gen3+ has demonstrated a 30% efficiency improvement over human labor in Tesla's factories, with future costs potentially dropping below $20,000 per unit, indicating both capability and affordability [2] Summary by Sections - **Order Acquisition** - Tesla's Optimus Gen3 has received a historic order of 10,000 units from PharmAGRI for automating drug production [1] - **Investment and Incentives** - Elon Musk's $10 billion stock purchase is tied to a compensation plan that rewards him significantly if Tesla achieves the delivery target of 1 million Optimus units [1] - **Efficiency and Cost** - The Optimus Gen3+ has been validated to be 30% more efficient than human workers, with a potential cost reduction to below $20,000, highlighting its economic viability [2]
具身智能能力狂飙,安全却滞后?首个安全可信EAI框架与路线图!
具身智能之心· 2025-09-18 00:03
编辑丨机器之心 为了弥合这一关键差距, 上海人工智能实验室和华东师范大学的研究团队 撰写了这篇 Position Paper,旨在为「安全可信具身智能」这一新兴领域建立一个系统性 的理论框架与发展蓝图,推动领域从碎片化研究走向整体性构建。 点击下方 卡片 ,关注" 具身智能之心 "公众号 >> 点击进入→ 具身 智能之心 技术交流群 更多干货,欢迎加入国内首个具身智能全栈学习社区 : 具身智能之心知识星球 (戳我) , 这里包含所有你想要的。 近年来,以人形机器人、自动驾驶为代表的具身人工智能(Embodied Artificial Intelligence, EAI)正以前所未有的速度发展,从数字世界大步迈向物理现实。然而, 当一次错误的风险不再是屏幕上的一行乱码,而是可能导致真实世界中的物理伤害时,一个紧迫的问题摆在了我们面前: 如何确保这些日益强大的具身智能体是安全且值得信赖的? 现实情况是,能力与安全,这两条本应齐头并进的轨道,正出现令人担忧的「脱钩」。如图 1 所示,业界的基础模型在能力上飞速迭代,却普遍忽视了与之匹配的 安全对齐机制;而学术界虽有探索,但研究成果往往零散、不成体系。 图 1: EA ...
TrajBooster:首个全身人行操作VLA方案,跨构型解决数据难题(代码全开源)
具身智能之心· 2025-09-18 00:03
Core Insights - The article discusses the TrajBooster framework, which aims to enhance the capabilities of humanoid robots by utilizing a trajectory-centric learning approach, enabling them to perform complex household tasks with minimal training data [2][40]. Group 1: Research Background and Challenges - The development of humanoid robots faces two main challenges: the unique difficulties of maintaining dynamic balance while performing upper body tasks, and the scarcity of high-quality training data necessary for effective VLA model training [3][4]. - Existing methods rely on expensive equipment and expert operators, resulting in limited data sets that do not adequately cover the diverse action spaces required for humanoid robots [4]. Group 2: TrajBooster Framework - TrajBooster utilizes a three-step process: real trajectory extraction, simulation redirection, and dual-stage fine-tuning, allowing for the conversion of extensive wheeled robot data into effective training resources for bipedal robots [5][40]. - The framework significantly reduces the dependency on costly data from similar robot types, enabling zero-shot skill transfer and improving the robustness and generalization of the VLA models [2][5]. Group 3: Methodology - The framework begins with extracting real trajectories from the Agibot-World Beta dataset, which contains over 1 million real robot trajectories, and then maps this data to the Unitree G1 robot's operational space [7][9]. - A hierarchical composite model is employed to decouple control into upper and lower body systems, enhancing the efficiency of whole-body manipulation [11][12]. Group 4: Experimental Results - TrajBooster demonstrated superior performance in various tasks, achieving the lowest position error (2.851 cm) and rotation error (6.231 degrees) in mobile scenarios, validating the advantages of hierarchical training and coordinated online DAgger [27]. - The framework's ability to adapt to unseen tasks was evidenced by its success in a "water transfer" task, which was not included in the training data, showcasing improved generalization capabilities [39][40]. Group 5: Limitations and Future Directions - The current implementation is limited by the precision of the Unitree Dex-3 hand, which only supports simple grasping tasks; future work will focus on integrating dexterous hands with tactile sensing for more complex manipulations [41]. - There is a need to address the visual input discrepancies and expand the framework to include mobile manipulation data, as the current research is primarily focused on static tasks [43][44].
3D/4D World Model(WM)近期发展的总结和思考
具身智能之心· 2025-09-18 00:03
Core Viewpoint - The article discusses the current state and future directions of embodied intelligence, particularly focusing on the development and optimization of 3D/4D world models, emphasizing the importance of data collection and utilization in training effective models [3][4]. Group 1: Current Research Focus - The majority of work in the first three quarters of the year has centered on data collection and utilization, specifically how to efficiently use video example data to train robust foundational models [3]. - There is a growing concern regarding the clarity and reliability of data collection methods, prompting a reevaluation of the approaches to data analysis and the development of 3D/4D world models [3][4]. Group 2: Approaches to 3D/4D World Models - Two main research approaches have emerged in the development of 3D/4D world models: implicit and explicit methods, each revealing limitations that have yet to be effectively addressed [4][7]. - Current research on explicit world models remains focused on static 3D scenes, with methods for constructing and enriching these scenes being well-established and ready for practical application [5]. Group 3: Challenges and Limitations - The existing methods for 3D geometry modeling, such as 3DGS, face challenges in surface optimization, leading to rough results despite attempts to improve through structured modifications [8]. - Issues related to lighting and surface quality in 3D reconstruction are being gradually optimized, but the overall design still faces significant hurdles, particularly in cross-physics simulator deployment [9]. Group 4: Future Directions - The article anticipates that future work will increasingly integrate physical knowledge into 3D/4D models, aiming to enhance the direct physical understanding and reasoning capabilities of models [15]. - There is an expectation for the emergence of new research that combines simulation and video generation to address existing gaps in the understanding of physical interactions and motion [14][15].
清华联手理想提出LightVLA:剪掉冗余token,推理速度提升38%!
具身智能之心· 2025-09-18 00:03
Core Insights - The article discusses the development of the LightVLA framework, which aims to enhance the efficiency and performance of Vision-Language-Action (VLA) models in robotics by addressing the computational redundancy associated with visual tokens [2][3]. Research Background and Core Challenges - VLA models are essential for embodied intelligence, converting visual information and language instructions into executable robot actions. However, they face a significant bottleneck due to the computational complexity that increases quadratically with the number of visual tokens [2]. - Existing optimization methods often compromise performance for efficiency, leading to the loss of critical semantic information [3]. Existing Optimization Limitations - Trade-off between efficiency and performance: Many token pruning methods sacrifice performance by retaining a fixed number of tokens [3]. - Incompatibility of pruning schemes: Current visual-language model pruning methods focus on global semantics, which does not translate well to VLA models that require local semantic attention [3]. - Poor deployment compatibility: Pruning methods based on attention scores are not adaptable to mainstream inference frameworks, limiting their practical application [3]. LightVLA Framework Design - LightVLA allows the model to autonomously learn to select task-relevant visual tokens through fine-tuning, rather than relying on manually set pruning ratios [4]. - The framework consists of three modules: visual encoder, LLM backbone, and action head, focusing solely on visual token pruning while retaining the [CLS] token for global information [4]. Core Methodology: Three-Stage Pruning Process 1. **Query Generation**: Task-oriented queries are generated to identify relevant visual tokens without introducing additional parameters [6]. 2. **Token Scoring**: Each visual token is scored based on its relevance to the task, with higher scores indicating stronger associations [10]. 3. **Token Selection**: A modified Gumbel-softmax approach is used for differentiable selection, allowing for end-to-end training of the pruning process [12]. Experimental Validation and Results Analysis - LightVLA demonstrated superior performance across various tasks in the LIBERO benchmark dataset, achieving an average success rate of 97.4%, which is a 2.9% improvement over the baseline model OpenVLA-OFT [16]. - The framework significantly reduces computational costs, achieving a 59.1% reduction in FLOPs and a 38.2% decrease in latency while maintaining high performance [18]. Ablation Studies and Qualitative Validation - The effectiveness of key design choices was confirmed through ablation studies, showing that the pruning process is task-oriented and dynamically adapts to the requirements of different tasks [20][24]. - LightVLA's pruning strategy focuses on retaining critical tokens related to the task while discarding redundant background tokens [24]. Comparison with MoE - LightVLA differs fundamentally from the Mixture of Experts (MoE) approach, as it prioritizes task performance by selecting visually relevant tokens, whereas MoE focuses on balancing expert load without emphasizing semantic relevance [28].
具身智能之心企业合作邀请函
具身智能之心· 2025-09-17 03:14
联系方式 添加商务微信oooops-life做进一步沟通。 具身智能之心是具身智能领域的优秀创作和宣传的媒体平台。近一年内,我们和多家具身公司签订长期合作事 项,包括但不限于产品宣传、品牌宣传、硬件代理、联合运营、教育产品研发等。 随着团队的不断扩大,我们期望在上述业务上和更多优秀的公司建立联系,推动具身领域的快速发展。欢迎有 相关业务需求的公司或团队联系我们。 我们期待进一步的合作!!! ...