MindVLA
Search documents
理想汽车-W(02015)3月交付新车41,053辆 同比增长11.94%
智通财经网· 2026-04-01 09:31
智通财经APP讯,理想汽车-W(02015)发布公告,2026年3月,理想汽车交付新车41,053辆,同比增长 11.94%。截至2026年3月31日,理想汽车历史累计交付量为1,635,357辆。 截至2026年3月31日,理想汽车在全国已有517家零售中心,覆盖160个城市;售后维修中心及授权服务中 心552家,覆盖223个城市。理想汽车在全国已投入使用4,057座理想超充站,拥有22,439个充电桩。 随着产能瓶颈的解决,3月,理想i6交付量超过2.4万辆。全新一代理想L9预计于2026年第二季度上市。 3月,公司在英伟达2026年GTC大会上发布下一代自动驾驶基础模型MindVLA以及3D ViT三维视觉编码 器,能够直接感知真实的三维物理世界,统一理解真实物理空间的几何和语义,向人类智能的空间理解 能力看齐。 ...
理想汽车-W(02015.HK)3月交付新车41,053辆 全新一代理想L9预计于2026年第二季度上市
Ge Long Hui· 2026-04-01 09:26
截至2026年3月31日,理想汽车在全国已有517家零售中心,覆盖160个城市;售后维修中心及授权服务 中心552家,覆盖223个城市。理想汽车在全国已投入使用4,057座理想超充站,拥有22,439个充电桩。 格隆汇4月1日丨理想汽车-W(02015.HK)宣布,2026年3月,理想汽车交付新车41,053辆。截至2026年3月 31日,理想汽车历史累计交付量为1,635,357辆。 随着产能瓶颈的解决,3月,理想i6交付量超过2.4万辆。全新一代理想L9预计于2026年第二季度上市。 3月,公司在英伟达2026年GTC大会上发布下一代自动驾驶基础模型MindVLA以及3D ViT三维视觉编码 器,能够直接感知真实的三维物理世界,统一理解真实物理空间的几何和语义,向人类智能的空间理解 能力看齐。 ...
Li Auto Inc. March 2026 Delivery Update
Globenewswire· 2026-04-01 08:30
BEIJING, China, April 01, 2026 (GLOBE NEWSWIRE) -- Li Auto Inc. (“Li Auto” or the “Company”) (Nasdaq: LI; HKEX: 2015), a leader in China’s new energy vehicle market, today announced that it delivered 41,053 vehicles in March 2026. As of March 31, 2026, Li Auto’s cumulative deliveries reached 1,635,357. With the production bottleneck resolved, Li i6 monthly deliveries surpassed 24,000 units in March. The all-new Li L9 is expected to launch in the second quarter of 2026. In March, at the NVIDIA GTC 2026, the ...
科技未来:视觉语言动作- 自动驾驶的下一个 AI 前沿-Future of Tech_ VLA as the next AI frontier in autonomous driving
2026-03-24 01:27
Summary of Key Points from the Conference Call Industry Overview - The focus is on the autonomous driving industry, particularly advancements in AI technologies and their implications for various automakers in Japan and China [1][4][5][12][13]. Core Insights and Arguments Global Autonomous Driving Penetration - Global L2+/L2++ penetration is projected to reach 36% by 2030, up from 15% in 2025, while L3 adoption is expected to be limited due to complexity and regulatory hurdles [1][25][28]. - In China, L2+/L2++ penetration is expected to rise to approximately 70% by 2030, significantly higher than the global average [30][34]. - The US market is anticipated to see L2+/L2++ penetration of around 36% by 2030, supported by consumer acceptance of advanced features [43][47]. Japan's Approach to Autonomous Driving - Japanese automakers are adopting varied strategies for commercialization, with Toyota leading through a 'multi-pathway' approach, combining internal development and partnerships [4][9][12]. - Upcoming models like Toyota's RAV4 and Sony Honda Mobility's Afeela are expected to drive the rollout of software-defined vehicles (SDVs) [8][12]. - The Japanese market is characterized by a cautious approach, prioritizing safety and reliability, with L2+/L2++ penetration projected at 29% by 2030 [48][49]. China's Competitive Landscape - Leading Chinese EV manufacturers such as XPeng and Li Auto are at the forefront of adopting Vision-Language-Action (VLA) models, enhancing user experience and decision-making capabilities [5][13]. - The intense competition among Chinese OEMs is accelerating the development of advanced driver-assistance systems (ADAS), which are becoming essential features in premium EVs [5][13]. - Concerns remain regarding the monetization potential of these technologies and the ability of Chinese OEMs to introduce advanced features in international markets [5][13]. Technological Shifts - The transition from rule-based systems to end-to-end (E2E) architectures is being driven by the need for faster deployment and improved handling of edge cases [2][9]. - VLA models are seen as the next frontier in E2E development, with companies like Waymo leveraging advanced AI to enhance navigation capabilities [3][9]. Additional Important Insights - Traditional auto parts suppliers face challenges as automakers assert more control over software layers, potentially reducing suppliers' revenue from design changes [11]. - Japan's government is promoting SDVs as a national priority, aiming for a 30% penetration target by 2030-2035, which may accelerate strategic initiatives across the sector [12]. - The role of high-definition (HD) maps remains relevant even in E2E systems, as they provide essential localization support and training data for AI models [66][67]. Investment Implications - Ratings for Japanese automakers include Outperform for Suzuki and Toyota, Market-Perform for Honda and Denso, and Underperform for Nissan, Mazda, and Subaru [12][14]. - In China, BYD and Xiaomi are rated as Outperform, while XPeng, NIO, and Li Auto are rated as Market-Perform [14]. This summary encapsulates the key points discussed in the conference call, highlighting the advancements and strategic directions of the autonomous driving industry in Japan and China.
深扒了学术界和工业界的「空间智能」,更多的还停留在表层......
自动驾驶之心· 2025-12-28 03:30
Core Viewpoint - The article emphasizes the transition of autonomous driving from "perception-driven" to "spatial intelligence" by 2025, highlighting the importance of understanding and interacting with the three-dimensional physical world [3]. Group 1: Spatial Intelligence Definition - Spatial intelligence is defined as the ability to perceive, represent, reason, decide, and interact with spatial information, which is crucial for the interaction between intelligent agents and the physical world [3]. - Current spatial intelligence is primarily focused on perception and representation, with significant room for improvement in reasoning, decision-making, and interaction capabilities [3]. Group 2: World Models and Simulation - GAIA-2 is a multi-view generative world model for autonomous driving that generates driving videos based on physical laws and conditions, addressing edge cases in driving scenarios [5]. - GAIA-3 enhances GAIA-2 by increasing the scale fivefold and capturing fine-grained spatiotemporal contexts, representing the physical causal structure of the real world [9]. - ReSim combines expert trajectories from the real world with simulated dangerous behaviors to achieve high-fidelity simulations of extreme driving scenarios [11]. Group 3: Multimodal Reasoning - The SIG framework introduces a structured graph scheme that encodes scene layouts and object relationships, aiming to enhance geometric reasoning in autonomous driving [16]. - OmniDrive generates a large-scale 3D question-answer dataset to align visual language models with 3D spatial understanding and planning [19]. - SimLingo addresses the alignment of driving behavior with semantic instructions through an action dreaming task, demonstrating the potential of general models in real-time decision-making [21]. Group 4: Real-time Digital Twins - DrivingRecon is a 4D Gaussian reconstruction model that predicts parameters from surround-view videos, enabling efficient dynamic scene reconstruction for autonomous driving [26]. - VR-Drive enhances robustness in driving systems by allowing real-time prediction of new viewpoints without scene optimization [29]. Group 5: Embodied Fusion - MiMo-Embodied is the first open-source cross-embodied model that integrates autonomous driving with embodied intelligence, showcasing significant transfer effects in spatial reasoning capabilities [31]. - DriveGPT4-V2 is a closed-loop end-to-end autonomous driving framework that outputs low-level control signals, evolving from visual understanding to closed-loop control [36]. Group 6: Industry Trends - By 2025, the industry is moving towards an end-to-end VLA architecture, leveraging large language models for driving decision-making [40]. - Waymo's EMMA model integrates multimodal inputs and outputs in a unified language space, enhancing complex reasoning in driving tasks [41]. - DeepRoute.ai's DeepRoute IO 2.0 architecture introduces chain-of-thought reasoning to address the "black box" issue in end-to-end models, improving user trust in autonomous systems [44].
L3自动驾驶量产元年,离L4的梦想又近了一步?
Xin Lang Cai Jing· 2025-12-17 06:30
Group 1 - The Ministry of Industry and Information Technology has approved the commercial operation of L3 autonomous driving for the first time in China, allowing vehicles to operate under specific conditions with the system taking over driving tasks [1] - The two models approved for L3 autonomous driving are Changan Deep Blue SL03 and Arcfox Alpha S6, marking a significant step towards the commercialization of L3 technology [1] - The year 2026 is anticipated to be the "mass production year" for L3 autonomous driving, with several companies aiming to launch L3 vehicles by then [3][4] Group 2 - The approval clarifies the responsibility division for L3 autonomous driving, indicating that if an accident occurs while the system is activated, the car manufacturer may bear primary responsibility [1] - The L3 level is seen as a crucial transition from "assisted driving" to "fully autonomous driving," with L4 expected to achieve greater breakthroughs [1][4] - Major automotive companies, including XPeng, Chery, and GAC, have set timelines for the mass production of L3 vehicles, with GAC planning to launch its first L3 model in Q4 of this year [3][4] Group 3 - The automotive industry is experiencing intensified competition in intelligent driving technologies, with companies like BYD, Geely, and Chery developing their own autonomous driving systems [9] - The integration of AI and data-driven technologies is becoming essential for enhancing autonomous driving capabilities, moving beyond traditional rule-based systems [9][12] - The VLA model is emerging as a key technology in the transition from L2 to L4 autonomous driving, offering improved scene reasoning and generalization capabilities [9][14] Group 4 - The shift towards L3 autonomous driving represents a new beginning for human-machine coexistence, with ongoing exploration in technology iteration and regulatory improvement [17] - Companies are increasingly focusing on in-house development of core technologies, such as battery technology and autonomous driving algorithms, to enhance brand competitiveness [16] - The balance between self-research and collaboration is crucial for companies to maintain technological leadership while managing costs [16][17]
以理想汽车为例,探寻自动驾驶的「大脑」进化史 - VLA 架构解析
自动驾驶之心· 2025-12-07 02:05
作者 | 我要吃鸡腿 编辑 | 自动驾驶之心 原文链接: https://zhuanlan.zhihu.com/p/1965839552158623077 点击下方 卡片 ,关注" 自动驾驶之心 "公众号 戳我-> 领取 自动驾驶近30个 方向 学习 路线 >>自动驾驶前沿信息获取 → 自动驾驶之心知识星球 本文只做学术分享,如有侵权,联系删文 在自动驾驶这个飞速迭代的领域,技术范式的更迭快得令人目不暇接。前年,行业言必称BEV(鸟瞰图视 角);去年,"端到端"(End-to-End)又成了新的技术高地。然而,每一种范式在解决旧问题的同时,似乎都 在催生新的挑战。 传统的"端到端"自动驾驶,即VA(Vision-Action,视觉-行动)模型,就暴露出一个深刻的矛盾:它就像一个 车技高超但沉默寡言的"老司机"。它能凭借海量数据训练出的"直觉",在复杂的路况中做出令人惊叹的丝滑操 作。但当您坐在副驾,心脏漏跳一拍后问它:"刚才为什么突然减速?"——它答不上来。 这就是"黑箱"问题:系统能"做对",但我们不知道它"为何做对"。这种无法解释、无法沟通的特性,带来了巨 大的信任危机。 自动驾驶的三大范式演进。(a) ...
李想:特斯拉V14也用了VLA相同的技术
自动驾驶之心· 2025-10-19 23:32
Core Insights - The article discusses the five stages of artificial intelligence (AI) as defined by OpenAI, emphasizing the importance of each stage in the development and application of AI technologies [17][18]. Group 1: Stages of AI Development - The first stage is Chatbots, which serve as a foundational model that compresses human knowledge, akin to a person completing their education [19][4]. - The second stage is Reasoners, which utilize supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF) to perform continuous reasoning tasks, similar to advanced academic training [20][21]. - The third stage is Agents, where AI begins to perform tasks autonomously, requiring a high level of professionalism and reliability, comparable to a person in a specialized job [22][23]. - The fourth stage is Innovators, focusing on the ability to generate and solve problems through real-world training and feedback, which is essential for enhancing the capabilities of AI [25][26]. - The fifth stage is Organizations, which manage multiple agents and innovations to prevent chaos, similar to how businesses manage human resources [27][28]. Group 2: Computational Needs - The demand for reasoning computational power is expected to increase by 100 times in the next five years, while training computational needs may expand by 10 times [10][29]. - The article highlights the necessity for both edge computing and cloud-based processing to support the various stages of AI development [28][29]. Group 3: Ideal Automotive Applications - The company is developing its own reasoning models (MindVLA/MindGPT) and agents (Driver Agent/Ideal Classmate Agent) to enhance its autonomous driving capabilities [31][33]. - By 2026, the company plans to equip its autonomous vehicles with self-developed advanced edge chips for deeper integration with AI [12][33]. Group 4: Training and Skill Development - Effective training for AI involves enhancing three key abilities: information processing, problem formulation and solving, and resource allocation [39][40][41]. - The article emphasizes that successful AI applications require extensive training, akin to the 10,000 hours of practice needed for mastery in a profession [36][42].
李想: 特斯拉V14也用了VLA相同技术|25年10月18日B站图文版压缩版
理想TOP2· 2025-10-18 16:03
Core Viewpoint - The article discusses the five stages of artificial intelligence (AI) as defined by OpenAI, emphasizing the importance of each stage in the development and application of AI technologies [10][11]. Group 1: Stages of AI - The first stage is Chatbots, which serve as a foundational model that compresses human knowledge, akin to a person completing their education [2][14]. - The second stage is Reasoners, which utilize supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF) to perform continuous reasoning tasks, similar to advanced academic training [3][16]. - The third stage is Agents, where AI begins to perform tasks autonomously, requiring a high level of reliability and professionalism, comparable to a person in a specialized job [4][17]. - The fourth stage is Innovators, focusing on generating and solving problems through reinforcement training, necessitating a world model for effective training [5][19]. - The fifth stage is Organizations, which manage multiple agents and innovations to prevent chaos, similar to corporate management [4][21]. Group 2: Computational Needs - The demand for reasoning computational power is expected to increase by 100 times, while training computational needs may expand by 10 times over the next five years [7][23]. - The article highlights the necessity for both edge and cloud computing to support the various stages of AI development, particularly in the Agent and Innovator phases [6][22]. Group 3: Ideal Self-Developed Technologies - The company is developing its own reasoning models (MindVLA/MindGPT), agents (Driver Agent/Ideal Classmate Agent), and world models to enhance its AI capabilities [8][24]. - By 2026, the company plans to equip its autonomous driving technology with self-developed advanced edge chips for deeper integration with AI [9][26]. Group 4: Training and Skill Development - The article emphasizes the importance of training in three key areas: information processing ability, problem formulation and solving ability, and resource allocation ability [33][36]. - It suggests that effective training requires real-world experience and feedback, akin to the 10,000-hour rule for mastering a profession [29][30].
理想基座模型负责人近期很满意的工作: RuscaRL
理想TOP2· 2025-10-03 09:55
Core Viewpoint - The article discusses the importance of reinforcement learning (RL) in enhancing the intelligence of large models, emphasizing the need for effective interaction between models and their environments to obtain high-quality feedback [1][2]. Summary by Sections Section 1: Importance of Reinforcement Learning - The article highlights that RL is crucial for the advancement of large model intelligence, with a focus on how to enable models to interact with broader environments to achieve capability generalization [1][8]. - It mentions various RL techniques such as RLHF (Reinforcement Learning from Human Feedback), RLAIF (AI Feedback Reinforcement Learning), and RLVR (Verifiable Reward Reinforcement Learning) as key areas of exploration [1][8]. Section 2: RuscaRL Framework - The RuscaRL framework is introduced as a solution to the exploration bottleneck in RL, utilizing educational psychology's scaffolding theory to enhance the reasoning capabilities of large language models (LLMs) [12][13]. - The framework employs explicit scaffolding and verifiable rewards to guide model training and improve response quality [13][15]. Section 3: Mechanisms of RuscaRL - **Explicit Scaffolding**: This mechanism provides structured guidance through rubrics, helping models generate diverse and high-quality responses while gradually reducing external support as the model's capabilities improve [14]. - **Verifiable Rewards**: RuscaRL designs rewards based on rubrics, allowing for stable and reliable feedback during training, which enhances exploration diversity and ensures knowledge consistency across tasks [15][16]. Section 4: Future Implications - The article suggests that both MindGPT and MindVLA, which target digital and physical worlds respectively, could benefit from the advancements made through RuscaRL, indicating a promising future for self-evolving models [9][10]. - It emphasizes that the current challenges in RL are not just algorithmic but also involve systemic integration of algorithms and infrastructure, highlighting the need for innovative approaches in building capabilities [9].