Group 1: Importance of Embodied AI and Multi-Sensor Fusion Perception - Embodied AI is a crucial direction in AI development, enabling autonomous decision-making and action through real-time perception in dynamic environments, with applications in autonomous driving and robotics [2][3] - Multi-sensor fusion perception (MSFP) is essential for robust perception and accurate decision-making in embodied AI, integrating data from various sensors like cameras, LiDAR, and radar to achieve comprehensive environmental awareness [2][3] Group 2: Limitations of Current Research - Existing AI-based MSFP methods have shown success in fields like autonomous driving but face inherent challenges in embodied AI, such as the heterogeneity of cross-modal data and temporal asynchrony between different sensors [3][4] - Current reviews on MSFP often focus on single tasks or research areas, limiting their applicability to researchers in related fields [4] Group 3: Overview of MSFP Research - The paper discusses the background of MSFP, including various perception tasks, sensor data types, popular datasets, and evaluation standards [5] - It reviews multi-modal fusion methods at different levels, including point-level, voxel-level, region-level, and multi-level fusion [5] Group 4: Sensor Data and Datasets - Various sensor data types are critical for perception tasks, including camera data, LiDAR data, and radar data, each with unique advantages and limitations [7][10] - The paper presents several datasets used in MSFP research, such as KITTI, nuScenes, and Waymo Open, detailing their characteristics and the types of data they provide [12][13][14] Group 5: Perception Tasks - Key perception tasks include object detection, semantic segmentation, depth estimation, and occupancy prediction, each contributing to the overall understanding of the environment [16][17] Group 6: Multi-Modal Fusion Methods - Multi-modal fusion methods are categorized into point-level, voxel-level, region-level, and multi-level fusion, each with specific techniques to enhance perception robustness [20][21][22][27] Group 7: Multi-Agent Fusion Methods - Collaborative perception techniques integrate data from multiple agents and infrastructure, addressing challenges like occlusion and sensor failures in complex environments [32][34] Group 8: Time Series Fusion - Time series fusion is a key component of MSFP systems, enhancing perception continuity across time and space, with methods categorized into dense, sparse, and hybrid queries [40][41] Group 9: Multi-Modal Large Language Model (MM-LLM) Fusion - MM-LLM fusion combines visual and textual data for complex tasks, with various methods designed to enhance the integration of perception, reasoning, and planning capabilities [53][54][57][59]
清华大学最新综述!当下智能驾驶中多传感器融合如何发展?
自动驾驶之心·2025-06-26 12:56