Autonomous Driving
Search documents
秋招正当时!自动驾驶之心求职交流群来啦~
自动驾驶之心· 2025-07-28 03:15
Group 1 - The article highlights the growing anxiety among job seekers, particularly students and professionals looking to transition into new fields, driven by the desire for better opportunities [1] - It notes that the landscape of autonomous driving technology is becoming more standardized, with a shift from numerous directions requiring algorithm engineers to a focus on unified models like one model, VLM, and VLA, indicating higher technical barriers [1] - The article emphasizes the importance of community building to support individuals in their career growth and industry knowledge, leading to the establishment of a job-related community for discussions on industry trends, company developments, and job opportunities [1]
传统感知和规控,打算转端到端VLA了...
自动驾驶之心· 2025-07-28 03:15
Core Viewpoint - The article emphasizes the shift in research focus from traditional perception and planning methods to end-to-end Vision-Language-Action (VLA) models in the autonomous driving field, highlighting the emergence of various subfields and the need for researchers to adapt to these changes [2][3]. Group 1: VLA Research Directions - The end-to-end development has led to the emergence of multiple technical subfields, categorized into one-stage and two-stage end-to-end approaches, with examples like PLUTO and UniAD [2]. - Traditional fields such as BEV perception and multi-sensor fusion are becoming mature, while the academic community is increasingly focusing on large models and VLA [2]. Group 2: Research Guidance and Support - The program offers structured guidance for students in VLA and autonomous driving, aiming to help them systematically grasp key theoretical knowledge and develop their own research ideas [7][10]. - The course includes a comprehensive curriculum covering classic and cutting-edge papers, coding implementation, and writing methodologies, ensuring students can produce a solid research paper [8][11]. Group 3: Enrollment and Requirements - The program is open to a limited number of students (6 to 8 per session) who are pursuing degrees in VLA and autonomous driving [6]. - Students are expected to have a foundational understanding of deep learning, Python, and PyTorch, with additional support provided for those needing to strengthen their basics [12][14]. Group 4: Course Structure and Outcomes - The course spans 12 weeks of online group research followed by 2 weeks of paper guidance, culminating in a maintenance period for the research paper [11]. - Participants will produce a draft of a research paper, receive project completion certificates, and may obtain recommendation letters based on their performance [15].
自驾一边是大量岗位,一遍是招不到人,太魔幻了......
自动驾驶之心· 2025-07-26 02:39
Core Viewpoint - The autonomous driving industry is experiencing a paradox where job vacancies exist alongside a scarcity of suitable talent, leading to a cautious hiring environment as companies prioritize financial sustainability and effective business models over rapid expansion [2][3]. Group 1: Industry Challenges - Many companies possess a seemingly complete technology stack (perception, control, prediction, mapping, data closure), yet they still face significant challenges in achieving large-scale, low-cost, and high-reliability commercialization [3]. - The gap between "laboratory results" and "real-world performance" remains substantial, indicating that practical application of technology is still a work in progress [3]. Group 2: Talent Acquisition - Companies are not necessarily unwilling to hire; rather, they have an unprecedented demand for "top talent" and "highly compatible talent" in the autonomous driving sector [4]. - The industry is shifting towards a more selective hiring process, focusing on candidates with strong technical skills and relevant experience in cutting-edge research and production [3][4]. Group 3: Community and Resources - The "Autonomous Driving Heart Knowledge Planet" is the largest community for autonomous driving technology in China, established to provide industry insights and facilitate talent development [9]. - The community has nearly 4,000 members and includes over 100 experts in the autonomous driving field, offering various learning pathways and resources [7][9]. Group 4: Learning and Development - The community emphasizes the importance of continuous learning and networking, providing a platform for newcomers to quickly gain knowledge and for experienced individuals to enhance their skills and connections [10]. - The platform includes comprehensive learning routes covering nearly all subfields of autonomous driving technology, such as perception, mapping, and AI model deployment [9][12].
Investor Reaction To Predictable Mobileye Earnings Was Negative: Analyst
Benzinga· 2025-07-25 18:34
Core Insights - Mobileye Global reported a fiscal second-quarter 2025 revenue of $506 million, a 15% year-on-year increase, surpassing analyst expectations of $463.26 million, with adjusted EPS of 13 cents exceeding the consensus estimate of 9 cents [1][3] - The company raised its fiscal 2025 revenue outlook to a range of $1.765 billion to $1.885 billion, up from the previous range of $1.690 billion to $1.810 billion, aligning closely with the analyst consensus estimate of $1.770 billion [2] Financial Performance - Shipments of approximately 9.7 million EyeQ units exceeded the analyst's estimate of 9 million, driven by strong demand from OEMs, particularly in China [6] - Adjusted gross margin for the quarter was 68.6%, slightly above the analyst's estimate of 68.4% and close to the Street's expectation of 68.8% [6] - Operating expenses were lower than anticipated at $241 million, resulting in adjusted operating income of $106 million, surpassing both expectations and the preliminary guidance [6] Future Outlook - Management emphasized 2027 as a critical year for revenue acceleration, driven by the adoption of SuperVision and initial deployments of Connected and Autonomous Vehicles (CAVs) [7] - Full-scale Drive deployments are planned for late 2026 across multiple U.S. and European cities, with the CAV business expected to contribute $150 million in 2027 revenue [8] - The company anticipates ADAS revenue could reach around $2 billion in 2027, which is considered a conservative estimate [8] Market Position and Partnerships - Mobileye's partnerships with major companies such as Volkswagen, Uber, and Lyft are expected to enhance its market position in the autonomous driving sector [9] - The company is transitioning to full production hardware for the ID. Buzz robotaxi, with teleoperations expected to begin in 2025 and driverless service planned for 2026 [9] Analyst Commentary - Needham analyst Quinn Bolton reiterated a buy rating on Mobileye with a price target of $18, despite the stock's decline following the earnings report [3][11] - Bolton noted that while management's tone was cautious, there is potential upside in fourth-quarter revenue and improving margin visibility, supporting a strong long-term growth trajectory for Mobileye [11]
小马智行开启7×24小时自动驾驶测试:「不眠模式」破解城市夜归难题
IPO早知道· 2025-07-25 13:15
Core Viewpoint - The article highlights the launch of 24/7 autonomous driving testing by Xiaoma Zhixing in Beijing, Guangzhou, and Shenzhen, marking a significant innovation in autonomous driving policies in these cities [2][4]. Group 1: Autonomous Driving Testing - Xiaoma Zhixing has expanded its testing hours from 7 AM to 11 PM to a full 24 hours, addressing the "forgotten hours" of late-night transportation where traditional services are limited [2]. - The company has accumulated over 50 million kilometers of autonomous driving testing mileage across various cities, demonstrating its extensive experience and capability in diverse conditions [4]. Group 2: Technology and Safety - The L4 autonomous driving system utilizes a multi-sensor fusion technology, including high-performance 128-line LiDAR and 8-megapixel cameras, ensuring real-time environmental recognition even in challenging low-light conditions [4][5]. - Xiaoma Zhixing's self-developed sensor cleaning solution effectively addresses perception accuracy issues caused by extreme weather conditions, enhancing driving safety [5]. Group 3: Urban Impact and Future Potential - The deployment of Xiaoma Zhixing's seventh-generation autonomous vehicles is reshaping urban operational logic, aiming to unlock the commercial potential and social value of autonomous driving services [7]. - The company envisions its 24/7 autonomous vehicles as "guardians" of the city, providing reliable transportation options for individuals during late-night hours [7].
传统的感知被嫌弃,VLA逐渐成为新秀......
自动驾驶之心· 2025-07-25 08:17
Core Insights - The article discusses the advancements in end-to-end autonomous driving algorithms, highlighting the emergence of various models and approaches in recent years, such as PLUTO, UniAD, OccWorld, and DiffusionDrive, which represent different technical directions in the field [1] - It emphasizes the shift in academic focus towards large models and Vision-Language-Action (VLA) methodologies, suggesting that traditional perception and planning tasks are becoming less prominent in top conferences [1] - The article encourages researchers to align their work with large models and VLA, indicating that there are still many subfields to explore despite the challenges for beginners [1] Summary by Sections Section 1: VLA Research Topics - The article introduces VLA research topics aimed at helping students systematically grasp key theoretical knowledge and expand their understanding of the specified direction [6] - It addresses the need for students to combine theoretical models with practical coding skills to develop new models and enhance their research capabilities [6] Section 2: Enrollment Information - The program has a limited enrollment capacity of 6 to 8 students per session [5] - It targets students at various academic levels (bachelor's, master's, and doctoral) who are interested in enhancing their research skills in autonomous driving and AI [7] Section 3: Course Outcomes - Participants will analyze classic and cutting-edge papers, understand key algorithms, and learn about writing and submission methods for academic papers [8][10] - The course includes a structured timeline of 12 weeks of online group research, followed by 2 weeks of paper guidance and a 10-week maintenance period [10] Section 4: Course Highlights - The program features a "2+1" teaching model with experienced instructors providing comprehensive support throughout the learning process [13] - It emphasizes high academic standards and aims to equip students with a rich set of outputs, including a paper draft and a project completion certificate [13] Section 5: Technical Requirements - Students are expected to have a foundational understanding of deep learning, basic programming skills in Python, and familiarity with PyTorch [11] - Hardware requirements include access to high-performance machines, preferably with multiple GPUs [11] Section 6: Service and Support - The program includes dedicated supervisors to track student progress and provide assistance with academic and non-academic issues [17] - The course will be conducted via Tencent Meeting and recorded for later access [18]
基于3DGS和Diffusion的自动驾驶闭环仿真论文总结
自动驾驶之心· 2025-07-24 09:42
Core Viewpoint - The article discusses advancements in autonomous driving simulation technology, highlighting the integration of various components such as scene rendering, data collection, and intelligent agents to create realistic driving environments [1][2][3]. Group 1: Simulation Components - The first step involves creating a static environment using 3D Gaussian Splatting and Diffusion Models to build a realistic cityscape, capturing intricate details [1]. - The second step focuses on data collection from panoramic views to extract dynamic assets like vehicles and pedestrians, enhancing the realism of simulations [2]. - The third step emphasizes relighting techniques to ensure that assets appear natural under various lighting conditions, simulating different times of day and weather scenarios [2]. Group 2: Intelligent Agents and Weather Systems - The fourth step introduces intelligent agents that mimic real-world behaviors, allowing for complex interactions within the simulation [3]. - The fifth step incorporates weather systems to enhance the atmospheric realism of the simulation, enabling scenarios like rain or fog [4]. Group 3: Advanced Features - The sixth step includes advanced features that challenge autonomous vehicles with unexpected obstacles, simulating real-world driving complexities [4].
端到端自动驾驶万字长文总结
自动驾驶之心· 2025-07-23 09:56
Core Viewpoint - The article discusses the current development status of end-to-end autonomous driving algorithms, comparing them with traditional algorithms and highlighting their advantages and limitations [1][3][53]. Summary by Sections Traditional vs. End-to-End Algorithms - Traditional autonomous driving algorithms follow a pipeline of perception, prediction, and planning, where each module has distinct inputs and outputs [3]. - End-to-end algorithms take raw sensor data as input and directly output path points, simplifying the process and reducing error accumulation [3][5]. - Traditional algorithms are easier to debug and have some level of interpretability, but they suffer from cumulative error issues due to the inability to ensure complete accuracy in perception and prediction modules [3][5]. Limitations of End-to-End Algorithms - End-to-end algorithms face challenges such as limited ability to handle corner cases, as they rely heavily on data-driven methods [7][8]. - The use of imitation learning in these algorithms can lead to difficulties in learning optimal ground truth and handling exceptional cases [53]. - Current end-to-end paradigms include imitation learning (behavior cloning and inverse reinforcement learning) and reinforcement learning, with evaluation methods categorized into open-loop and closed-loop [8]. Current Implementations - The ST-P3 algorithm is highlighted as an early work focusing on end-to-end autonomous driving, utilizing a framework that includes perception, prediction, and planning modules [10][11]. - Innovations in the ST-P3 algorithm include a perception module that uses a self-centered cumulative alignment technique and a prediction module that employs a dual-path prediction mechanism [11][13]. - The planning phase of ST-P3 optimizes predicted trajectories by incorporating traffic light information [14][15]. Advanced Techniques - The UniAD system employs a full Transformer framework for end-to-end autonomous driving, integrating multiple tasks to enhance performance [23][25]. - The TrackFormer framework focuses on the collaborative updating of track queries and detect queries to improve prediction accuracy [26]. - The VAD (Vectorized Autonomous Driving) method introduces vectorized representations for better structural information and faster computation in trajectory planning [32][33]. Future Directions - The article suggests that end-to-end algorithms still primarily rely on imitation learning frameworks, which have inherent limitations that need further exploration [53]. - The introduction of more constraints and multi-modal planning methods aims to address trajectory prediction instability and improve model performance [49][52].
聊聊自动驾驶闭环仿真和3DGS!
自动驾驶之心· 2025-07-22 12:46
Core Viewpoint - The article discusses the development and implementation of the Street Gaussians algorithm, which aims to efficiently model dynamic street scenes for autonomous driving simulations, addressing previous limitations in training and rendering speeds [2][3]. Group 1: Background and Challenges - Previous methods faced challenges such as slow training and rendering speeds, as well as inaccuracies in vehicle pose tracking [3]. - The Street Gaussians algorithm represents dynamic urban street scenes as a combination of point-based backgrounds and foreground objects, utilizing optimized vehicle tracking poses [3][4]. Group 2: Technical Implementation - The background model is represented as a set of points in world coordinates, each assigned a 3D Gaussian to depict geometric shape and color, with parameters including covariance matrices and position vectors [8]. - The object model for moving vehicles includes a set of optimizable tracking poses and point clouds, with similar Gaussian attributes to the background model but defined in local coordinates [11]. Group 3: Innovations in Appearance Modeling - The article introduces a 4D spherical harmonic model to encode temporal information into the appearance of moving vehicles, reducing storage costs compared to traditional methods [12]. - The effectiveness of the 4D spherical harmonic model is demonstrated, showing significant improvements in rendering results and reducing artifacts [16]. Group 4: Initialization Techniques - Street Gaussians utilizes aggregated LiDAR point clouds for initialization, addressing the limitations of traditional SfM point clouds in urban environments [17]. Group 5: Course and Learning Opportunities - The article promotes a specialized course on 3D Gaussian Splatting (3DGS), covering various subfields and practical applications in autonomous driving, aimed at enhancing understanding and implementation skills [26][30].
8万条!清华开源VLA数据集:面向自动驾驶极端场景,安全提升35%
自动驾驶之心· 2025-07-22 12:46
Core Viewpoint - The article discusses the development of the Impromptu VLA dataset, which aims to address the data scarcity issue in unstructured driving environments for autonomous driving systems. It highlights the dataset's potential to enhance the performance of vision-language-action models in complex scenarios [4][29]. Dataset Overview - The Impromptu VLA dataset consists of over 80,000 meticulously constructed video clips, extracted from more than 2 million original materials across eight diverse open-source datasets [5][29]. - The dataset focuses on four key unstructured challenges: boundary-ambiguous roads, temporary traffic rule changes, unconventional dynamic obstacles, and complex road conditions [12][13]. Methodology - The dataset construction involved a multi-step process, including data collection, scene classification, and multi-task annotation generation, utilizing advanced visual-language models (VLMs) for scene understanding [10][17]. - A rigorous manual verification process was implemented to ensure high-quality annotations, with significant F1 scores achieved for various categories, confirming the reliability of the VLM-based annotation process [18]. Experimental Validation - The effectiveness of the Impromptu VLA dataset was validated through comprehensive experiments, showing significant performance improvements in mainstream autonomous driving benchmarks. For instance, the average score in the closed-loop NeuroNCAP test improved from 1.77 to 2.15, with a reduction in collision rates from 72.5% to 65.5% [6][21]. - In open-loop trajectory prediction evaluations, models trained with the Impromptu VLA dataset achieved L2 errors as low as 0.30 meters, demonstrating competitive performance compared to leading methods that rely on larger proprietary datasets [24]. Conclusion - The Impromptu VLA dataset serves as a critical resource for developing more robust and adaptive autonomous driving systems capable of handling complex real-world scenarios. The research confirms the dataset's significant value in enhancing perception, prediction, and planning capabilities in unstructured driving environments [29].