Core Insights - Ant Group's Lingbo Technology has made significant advancements in spatial intelligence by open-sourcing the high-precision spatial perception model LingBot-Depth, which enhances depth perception and 3D spatial understanding for robots and autonomous vehicles [1] Group 1: Model Performance - LingBot-Depth demonstrates a generational advantage in authoritative benchmark evaluations, reducing relative error (REL) by over 70% compared to mainstream models like PromptDA and PriorDA in indoor scenes, and achieving a 47% reduction in RMSE error in challenging sparse SfM tasks [1] - The model excels in handling transparent and reflective objects, which are common in household and industrial environments, overcoming limitations faced by traditional depth cameras [1][2] Group 2: Technology and Innovation - The "Masked Depth Modeling" (MDM) technology developed by Lingbo Technology allows the model to infer and complete missing depth data by integrating texture, contours, and contextual information from RGB images, resulting in clearer and more complete 3D depth maps [2] - LingBot-Depth has been certified by the Oubo Zhongguang Depth Vision Laboratory, achieving industry-leading levels in accuracy, stability, and adaptability to complex scenes [2] Group 3: Data and Collaboration - The model's superiority is attributed to a vast dataset, with approximately 10 million raw samples and 2 million high-value depth pair data used for training, which will soon be open-sourced to accelerate community efforts in tackling complex spatial perception challenges [3] - Ant Group's Lingbo Technology has reached a strategic cooperation intention with Oubo Zhongguang to launch a new generation of depth cameras based on LingBot-Depth's capabilities [3]
让机器人“看清”三维世界,蚂蚁灵波开源空间感知模型