Workflow
3D Gaussian Splatting (3DGS)
icon
Search documents
ICCV 2025自动驾驶场景重建工作汇总!这个方向大有可为~
自动驾驶之心· 2025-07-29 00:52
Core Viewpoint - The article emphasizes the advancements in autonomous driving scene reconstruction, highlighting the integration of various technologies and the collaboration among top universities and research institutions in this field [2][12]. Summary by Sections Section 1: Overview of Autonomous Driving Scene Reconstruction - The article discusses the importance of dynamic and static scene reconstruction in autonomous driving, focusing on the need for precise color and geometric information through the integration of lidar and visual data [2]. Section 2: Research Contributions - Several notable research works from prestigious institutions such as Tsinghua University, Nankai University, Fudan University, and the University of Illinois Urbana-Champaign are mentioned, showcasing their contributions to the field [5][6][10][11]. Section 3: Educational Initiatives - The article promotes a comprehensive course on 3D Gaussian Splatting (3DGS), designed in collaboration with leading experts, aimed at providing in-depth knowledge and practical skills in autonomous driving scene reconstruction [15][19]. Section 4: Course Structure - The course is structured into eight chapters, covering foundational algorithms, technical details of 3DGS, static and dynamic scene reconstruction, surface reconstruction, and practical applications in autonomous driving [19][21][23][25][27][29][31][33]. Section 5: Target Audience - The course is targeted at researchers, students, and professionals interested in 3D reconstruction, requiring a foundational understanding of 3DGS and related technologies [36][37].
多样化大规模数据集!SceneSplat++:首个基于3DGS的综合基准~
自动驾驶之心· 2025-06-20 14:06
Core Insights - The article introduces SceneSplat-Bench, a comprehensive benchmark for evaluating visual-language scene understanding methods based on 3D Gaussian Splatting (3DGS) [11][30]. - It presents SceneSplat-49K, a large-scale dataset containing approximately 49,000 raw scenes and 46,000 filtered 3DGS scenes, which is the most extensive open-source dataset for complex and high-quality scene-level 3DGS reconstruction [9][30]. - The evaluation indicates that generalizable methods consistently outperform per-scene optimization methods, establishing a new paradigm for scalable scene understanding through pre-trained models [30]. Evaluation Protocols - The benchmark evaluates methods based on two key metrics in 3D space: foreground mean Intersection over Union (f-mIoU) and foreground mean accuracy (f-mAcc), addressing object size imbalance and reducing viewpoint dependency compared to 2D evaluations [22][30]. - The evaluation dataset includes ScanNet, ScanNet++, and Matterport3D for indoor scenes, and HoliCity for outdoor scenes, emphasizing the methods' capabilities across various object scales and complex environments [22][30]. Dataset Contributions - SceneSplat-49K is compiled from multiple sources, including SceneSplat-7K, DL3DV-10K, HoliCity, and Aria Synthetic Environments, ensuring a diverse range of indoor and outdoor environments [9][10]. - The dataset preparation involved approximately 891 GPU days and extensive human effort, highlighting the significant resources invested in creating a high-quality dataset [7][9]. Methodological Insights - The article categorizes methods into three types: per-scene optimization methods, per-scene optimization-free methods, and generalizable methods, with SceneSplat representing the latter [23][30]. - Generalizable methods eliminate the need for extensive single-scene computations during inference, allowing for efficient processing of 3D scenes in a single forward pass [24][30]. Performance Results - The results from SceneSplat-Bench demonstrate that SceneSplat excels in both performance and efficiency, often surpassing the pseudo-label methods used for its pre-training [24][30]. - The performance of various methods shows significant variation based on the dataset's complexity, indicating the importance of challenging benchmarks in revealing the limitations of competing methods [28][30].