Workflow
新视角合成
icon
Search documents
随手拍照片就能VR云旅游!无位姿、稀疏图像条件下实现稳定3D重建和新视角合成|港科广
量子位· 2025-07-31 04:23
Core Viewpoint - A new algorithm, RegGS, developed by the Hong Kong University of Science and Technology (Guangzhou), can reconstruct 3D models from sparse 2D images without precise camera positioning, achieving centimeter-level accuracy suitable for VR applications [2][4]. Group 1: Methodology - RegGS combines feed-forward Gaussian representation with structural registration to address the challenges of sparse and pose-less images, providing a new pathway for practical 3D reconstruction [6][8]. - The core mechanism involves registering local 3D Gaussian mixture models to gradually build a global 3D scene, avoiding reliance on traditional Structure from Motion (SfM) initialization and requiring fewer input images [8][12]. Group 2: Experimental Results - In experiments on the RE10K and ACID datasets, RegGS outperformed existing mainstream methods across various input frame counts (2×/8×/16×/32×) in metrics such as PSNR, SSIM, and LPIPS [9][12]. Group 3: Applications - RegGS addresses the "sparse + pose-less" problem with significant real-world applications, including: - 3D reconstruction from user-generated content (UGC) videos, which often lack camera parameters [13]. - Drone aerial mapping, demonstrating robustness to large viewpoint variations and low frame rates [13]. - Restoration of historical images/documents, enabling 3D reconstruction from a few photos taken from different angles [13]. - Compared to traditional SfM or Bundle Adjustment methods, RegGS requires less structural input and is more feasible for unstructured data applications [13]. Group 4: Limitations and Future Directions - The performance and efficiency of RegGS are currently limited by the quality of the upstream feed-forward model and the computational cost of the MW2 distance calculation, indicating areas for future optimization [13].
暑假打比赛!RealADSim Workshop智驾挑战赛正式开启,奖池总金额超30万(ICCV'25)
自动驾驶之心· 2025-07-11 09:42
Core Viewpoint - The article emphasizes the significance of high-fidelity simulation technology in overcoming the challenges of testing autonomous driving algorithms, particularly through the introduction of New View Synthesis (NVS) technology, which allows for the creation of closed-loop driving simulation environments based on real-world data [1][2]. Group 1: Challenges and Tasks - The workshop addresses two main challenges in the application of NVS technology, focusing on the need for improved rendering quality in extrapolated views and the evaluation of driving algorithms in closed-loop simulation environments [2][3]. - The first track, "Extrapolated View New View Synthesis," aims to enhance rendering quality under sparse input views, which is crucial for evaluating autonomous driving algorithms in various scenarios [3][4]. - The second track, "Closed-Loop Simulation Evaluation," highlights the importance of creating high-fidelity simulation environments that bridge the gap between real-world data and interactive assessments, overcoming the limitations of traditional static datasets [5][6]. Group 2: Competition Details - Each track of the workshop offers awards, including a Creative Award of $9,000, and the competition is set to commence on June 30, 2025, with submissions due by August 31, 2025 [8][9]. - The workshop encourages global participation to advance autonomous driving technology, providing a platform for challenging and valuable research [10][11].