Workflow
特斯拉FSD V14
icon
Search documents
李想: 特斯拉V14也用了VLA相同技术|25年10月18日B站图文版压缩版
理想TOP2· 2025-10-18 16:03
Core Viewpoint - The article discusses the five stages of artificial intelligence (AI) as defined by OpenAI, emphasizing the importance of each stage in the development and application of AI technologies [10][11]. Group 1: Stages of AI - The first stage is Chatbots, which serve as a foundational model that compresses human knowledge, akin to a person completing their education [2][14]. - The second stage is Reasoners, which utilize supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF) to perform continuous reasoning tasks, similar to advanced academic training [3][16]. - The third stage is Agents, where AI begins to perform tasks autonomously, requiring a high level of reliability and professionalism, comparable to a person in a specialized job [4][17]. - The fourth stage is Innovators, focusing on generating and solving problems through reinforcement training, necessitating a world model for effective training [5][19]. - The fifth stage is Organizations, which manage multiple agents and innovations to prevent chaos, similar to corporate management [4][21]. Group 2: Computational Needs - The demand for reasoning computational power is expected to increase by 100 times, while training computational needs may expand by 10 times over the next five years [7][23]. - The article highlights the necessity for both edge and cloud computing to support the various stages of AI development, particularly in the Agent and Innovator phases [6][22]. Group 3: Ideal Self-Developed Technologies - The company is developing its own reasoning models (MindVLA/MindGPT), agents (Driver Agent/Ideal Classmate Agent), and world models to enhance its AI capabilities [8][24]. - By 2026, the company plans to equip its autonomous driving technology with self-developed advanced edge chips for deeper integration with AI [9][26]. Group 4: Training and Skill Development - The article emphasizes the importance of training in three key areas: information processing ability, problem formulation and solving ability, and resource allocation ability [33][36]. - It suggests that effective training requires real-world experience and feedback, akin to the 10,000-hour rule for mastering a profession [29][30].
揭秘特斯拉FSD V14 “车位到车位”核心算法:高保真3D Occ占用预测
自动驾驶之心· 2025-10-11 16:03
Core Insights - The article discusses Tesla's FSD V14 and its innovative "space occupancy detection" algorithm, which allows for high-precision 3D spatial reconstruction using only 2D image data from cameras, achieving accuracy within 10 cm [4][11][20]. Group 1: Overview of the High-Fidelity 3D Occupancy Algorithm - The high-fidelity 3D occupancy algorithm utilizes AI to accurately perceive and make decisions in complex dynamic environments, focusing on the occupancy attributes of surrounding space [5][6]. - Key components of the algorithm include the occupancy grid algorithm, which predicts the occupancy status of voxels (3D pixels) around the vehicle [5][6]. Group 2: Technical Mechanisms - The algorithm employs a Signed Distance Function (SDF) to predict the distance to the nearest occupied voxel, enhancing spatial perception and enabling more refined shape recognition [7][18]. - The system processes images from multiple cameras using convolutional neural networks (CNN) to extract meaningful features, which are then transformed into 3D spatial representations [12][20]. Group 3: Applications and Use Cases - The high-fidelity occupancy network can be applied in advanced parking assistance systems, enabling the identification of available parking spaces and assessing their suitability based on various factors [23][24]. - The algorithm is also applicable in autonomous robots for indoor navigation, allowing them to distinguish between obstacles and navigable areas [29]. Group 4: Advantages and Innovations - The SDF-based rendering approach provides richer detail and smoother visuals compared to traditional point cloud or binary voxel occupancy rendering methods [21]. - The algorithm's reliance solely on 2D visual data, without the need for depth cameras or LiDAR, represents a significant innovation in the field of autonomous driving [11][12].