机器人导航
Search documents
刚刚,UCLA周博磊也加入了一家机器人公司
机器之心· 2025-10-15 02:54
Core Insights - Coco Robotics has appointed Bolei Zhou, a UCLA associate professor, as the Chief AI Scientist to lead the newly established Physical AI Lab, focusing on autonomous driving for sidewalks [1][2][4] - The company aims to achieve full automation in last-mile delivery, leveraging the extensive operational data collected over the past five years [2][4][6] Group 1: Company Overview - Coco Robotics, founded in 2020, specializes in last-mile delivery robotics and initially relied on teleoperators to navigate obstacles [2][4] - The company has accumulated millions of miles of data in complex urban environments, which is crucial for training reliable AI systems [4][6] Group 2: Research and Development - The Physical AI Lab will utilize the data collected to enhance automation and operational efficiency, focusing on local models used by their robots [6] - The lab operates independently from Coco Robotics' collaboration with OpenAI, which allows the use of OpenAI's models while sharing data for mutual benefit [5][6] Group 3: Bolei Zhou's Background - Bolei Zhou holds a PhD from MIT and has a strong research background in machine perception and intelligent decision-making, with over 100 publications and significant contributions to explainable AI [9][11][13] - His notable works include Class Activation Mapping (CAM) and the creation of the Places database, which contains over 10 million labeled scene images, enhancing scene recognition capabilities [11][18][20]
宇树科技公布导航专利,提升机器人巡检
Xin Lang Ke Ji· 2025-08-22 09:30
Core Viewpoint - Hangzhou Yushu Technology Co., Ltd. has recently published a patent for a "robot navigation interaction control method and system based on multi-map fusion," aimed at improving cross-scenario navigation for robots [1]. Group 1: Patent Details - The patent addresses the limitations of existing robot navigation solutions, which lack effective cross-scenario navigation mechanisms, resulting in inaccurate navigation across different areas [1]. - The invention utilizes the topological relationships between scenes to fuse multiple point cloud files, creating a comprehensive cross-scenario map [1]. - Based on this cross-scenario map and navigation tasks, the system performs path planning to generate navigation data, leading to a more efficient, reliable, and user-friendly cross-scenario navigation mechanism [1]. Group 2: Impact on Robotics - The proposed method enhances the ability of robots to navigate in complex environments or areas with significant scene differences, thereby improving operational inspection efficiency [1].
有几个Top具身公司的大模型、强化学习、VLA和具身导航岗位!
具身智能之心· 2025-07-10 03:36
Core Viewpoint - The article discusses job opportunities in the fields of multimodal large models, reinforcement learning, and navigation, highlighting positions in a unicorn company with ample funding [1]. Group 1: Multimodal Large Models - Job locations are in Beijing and Shenzhen with a salary range of 40k-80k/month [2]. - Responsibilities include developing cutting-edge algorithms for embodied intelligent multimodal large models applicable in various indoor and outdoor scenarios, focusing on framework design, model optimization, and training for navigation and operation tasks [2]. - Candidates should have a master's degree or higher in computer science, artificial intelligence, robotics, or control engineering, along with extensive experience in robot perception, navigation, and AI large models [3]. - Preferred qualifications include experience with algorithms related to multimodal large models in robot navigation and a solid foundation in algorithm development and engineering implementation [3][4]. Group 2: Reinforcement Learning - Job location is in Beijing with a salary range of 40k-80k/month [5]. - Specific job descriptions and requirements are not detailed in the provided text [5]. Group 3: Embodied Navigation Algorithms - Job location is in Shenzhen with a salary range of 30k-60k/month [6]. - The role involves researching and developing algorithms for embodied intelligence, focusing on the integration of multimodal data into planning and achieving end-to-end mapping from data to actions [6]. Group 4: Additional Qualifications - Candidates should have a strong foundation in machine learning, deep learning, and reinforcement learning, with the ability to conduct independent research in embodied intelligence and related fields [7]. - Experience in publishing papers in top conferences and journals is a plus, along with strong coding skills and participation in robotics competitions [7].
我在哪?要去哪?要怎么去?字节跳动提出Astra双模型架构助力机器人自由导航
机器之心· 2025-06-23 09:39
Core Viewpoint - The article discusses the challenges faced by traditional navigation systems in mobile robotics and introduces ByteDance's innovative dual-model architecture, Astra, which aims to enhance navigation capabilities in complex indoor environments [2][4]. Group 1: Traditional Navigation Challenges - Mobile robots must address three core navigation challenges: goal localization, self-localization, and path planning, which are critical for safe and reliable movement in complex environments [3]. - Traditional navigation systems often rely on multiple modules and small models, which can be inefficient and require further exploration for effective integration [3]. Group 2: Astra Dual-Model Architecture - Astra consists of two sub-models: Astra-Global for low-frequency tasks like goal and self-localization, and Astra-Local for high-frequency tasks such as local path planning and odometry estimation [5]. - Astra-Global utilizes a multimodal large language model (MLLM) to process visual and language inputs for precise localization on a global map [8][10]. Group 3: Astra-Global Functionality - Astra-Global employs a two-stage process for visual-language localization, achieving high accuracy in identifying locations based on visual inputs and natural language instructions [11][12]. - The model's training involves diverse datasets and a reward-based optimization approach, resulting in a significant improvement in localization accuracy, achieving 99.9% in unseen environments compared to 93.7% with traditional methods [12]. Group 4: Astra-Local Functionality - Astra-Local is designed for efficient local path generation and odometry estimation, incorporating a 4D spatiotemporal encoder and a planning head [13][15]. - The planning head utilizes a transformer-based flow matching method to generate executable trajectories while minimizing collision rates through a mask ESDF loss approach [16][23]. Group 5: Experimental Validation - Extensive experiments in various indoor environments, including warehouses and offices, validate Astra's innovative architecture and algorithm effectiveness [19]. - Astra-Global demonstrates superior multimodal localization capabilities, significantly outperforming traditional visual place recognition methods in accuracy and robustness [20][23]. Group 6: Future Prospects - Astra has potential applications in diverse environments such as shopping malls, hospitals, and libraries, enhancing service efficiency and user experience [25]. - Future improvements are planned for Astra-Global's semantic detail retention and the introduction of active exploration mechanisms to enhance localization robustness in complex settings [25][26].
还不知道发什么方向论文?别人已经投稿CCF-A了......
具身智能之心· 2025-06-18 03:03
Group 1 - The core viewpoint of the article is the launch of a mentoring program for students aiming to publish papers in top conferences such as CVPR and ICRA, building on last year's successful outcomes [1] - The mentoring directions include multimodal large models, VLA, robot navigation, robot grasping, embodied generalization, embodied synthetic data, end-to-end embodied intelligence, and 3DGS [2] - The mentors have published papers in top conferences like CVPR, ICCV, ECCV, ICLR, RSS, ICML, and ICRA, indicating their rich guiding experience [3] Group 2 - Students are required to submit a resume and must come from a domestic top 100 university or an international university ranked within QS 200 [4][5]