Workflow
强化学习
icon
Search documents
亲身体验后,我们总结了全网首份AI语音输入法红黑榜|锦秋AI实验室
锦秋集· 2026-01-08 14:57
Core Viewpoint - The article evaluates seven AI voice input tools across five real-life scenarios to determine which can accurately transcribe spoken language into written text, focusing on their ability to maintain the integrity of the original message and avoid misinterpretations [1][3][36]. Group 1: Evaluation Methodology - The evaluation involves a comparative analysis of seven AI voice input tools, assessing their performance in various scenarios such as casual conversations and meeting minutes [2][6]. - The assessment criteria include text consistency, local quality (handling of homophones, numbers, and punctuation), and overall user experience [11][28]. Group 2: Performance Results - In the first round of testing, tools like Sogou, iFly, and Doubao demonstrated high accuracy in transcribing casual conversations, maintaining key information effectively [10][19]. - The second round focused on meeting minutes, where Typeless and Doubao excelled in structuring information clearly, while Sogou and iFly struggled with critical errors that could lead to miscommunication [17][25]. - The third round tested mixed-language input, revealing that Doubao and Typeless maintained accuracy in technical terms, while Sogou and iFly faced significant issues with misinterpretation [26][30]. Group 3: Key Findings - The analysis concluded that the ability to accurately transcribe and maintain the original meaning of spoken language is crucial, especially in professional settings where errors can lead to serious misunderstandings [36]. - Typeless emerged as the top choice for structured documentation, while Doubao was recognized for its overall reliability in various scenarios [38][40]. - Tools like Sogou and Flash Talk were deemed unsuitable for high-stakes environments due to frequent critical errors that could compromise communication [40].
当我们把端到端量产需要的能力展开后......
自动驾驶之心· 2026-01-08 09:07
Core Viewpoint - The article emphasizes the rising importance of end-to-end (E2E) systems in the autonomous driving industry, highlighting the shift from modular perception to direct environmental sensing and action generation, which simplifies system complexity and enhances the ability to handle complex driving scenarios [2]. Group 1: End-to-End Systems - The success of Horizon HSD has prompted a reevaluation of the significance of E2E systems in smart driving, moving away from heavy reliance on modular perception and strict rule-based systems [2]. - E2E systems face challenges in practical applications, such as trajectory instability, primarily due to the lack of continuous correction capabilities based on environmental feedback [3]. - Reinforcement Learning (RL) offers a new approach for E2E systems, transitioning from imitation to optimization by incorporating reward signals to refine action strategies and address limitations of pure imitation learning [4][5]. Group 2: Industry Trends and Talent Demand - Leading companies in the industry have developed a comprehensive model iteration approach, which includes imitation learning training, closed-loop reinforcement learning, and rule-based planning, indicating a high barrier to entry for talent in E2E production [6]. - The high barrier to entry and scarcity of skilled professionals have resulted in generous salaries, with top talents earning starting salaries of 1 million and above [7]. Group 3: Challenges in Mass Production - The mass production of E2E systems encounters numerous challenges, including complex scenarios like congestion, static yaw, and collision situations, necessitating both data mining and data cleaning [8]. - There is a notable gap in practical experience among many candidates, as many have only theoretical knowledge without real-world application experience [8]. Group 4: Course Offering - The article introduces a specialized course aimed at bridging the gap in practical skills for E2E systems, led by top-tier algorithm engineers from the industry [9]. - The course covers various aspects of E2E systems, including task overview, two-stage and one-stage algorithms, navigation information applications, RL algorithms, trajectory optimization, and production experiences [12][14][15][16][17][18][19][20][21]. Group 5: Target Audience and Prerequisites - The course is designed for advanced learners with a foundational understanding of autonomous driving perception, reinforcement learning, and programming skills, although those with weaker backgrounds can still participate [22][23].
随到随学!端到端与VLA自动驾驶小班课(视频+答疑)
自动驾驶之心· 2026-01-08 05:58
Core Viewpoint - The article discusses an advanced course on end-to-end (E2E) autonomous driving, focusing on the latest technologies such as BEV perception, Visual Language Models (VLM), diffusion models, and reinforcement learning, aimed at equipping participants with cutting-edge skills in the field [1][4][8]. Group 1: Course Structure - The course is divided into several chapters, starting with an introduction to end-to-end algorithms, covering the historical development and advantages of E2E methods over modular approaches [4]. - The second chapter focuses on background knowledge essential for understanding E2E technologies, including VLA, diffusion models, and reinforcement learning, which are crucial for job interviews in the next two years [5][9]. - The third chapter delves into two-stage E2E methods, discussing their emergence, advantages, and notable algorithms like PLUTO and CarPlanner [5][6]. - The fourth chapter highlights one-stage E2E methods and VLA, exploring various subfields and their contributions to achieving the ultimate goals of E2E systems [6][10]. Group 2: Practical Application - The course includes a major project on RLHF fine-tuning, allowing participants to apply their knowledge in practical scenarios, including building pre-training and reinforcement learning modules [7]. - The course aims to help participants reach a level equivalent to one year of experience as an E2E autonomous driving algorithm engineer, covering various methodologies and key technologies [13]. Group 3: Target Audience and Requirements - The course is designed for individuals with a foundational understanding of autonomous driving, familiar with basic modules, and concepts like transformer models, reinforcement learning, and BEV perception [11]. - Participants are expected to have a background in probability theory and linear algebra, as well as proficiency in Python and PyTorch [11].
人工智能测评初创企业LMArena新一轮融资后估值达17亿美元
Xin Lang Cai Jing· 2026-01-07 09:30
公司首席执行官兼联合创始人阿纳斯塔西奥斯・安杰洛普洛斯表示:"头部人工智能实验室选择与我们 合作,是因为他们很难自行判断旗下模型的优劣水平。" 不过,部分模型开发商对 LMArena 的测评方式提出了质疑。他们认为,依靠无偿互联网用户提供反馈 的模式存在缺陷,不仅容易被人为操纵,而且结果也无法反映行业专家的专业意见。与之形成对比的 是,LMArena 的竞争对手 —— 例如数据标注初创企业 Scale AI—— 会聘请律师、教授等专业人士对模 型进行评估打分。 针对外界的质疑,LMArena 回应称,普通用户在评判与自身相关的问题答案时,往往具备更精准的判 断力,而且不向专家支付报酬的模式,能够让平台获取更真实客观的反馈。 LMArena 联合创始人阿 纳斯塔西奥斯・安杰洛普洛斯 据 LMArena 公司透露,这家凭借人工智能模型性能排名体系广受行业认可的初创企业,在新一轮融资 中筹集到 1.5 亿美元资金,投后估值(含本次融资额)达到 17 亿美元。这一估值较 2025 年 5 月披露的 种子轮融资估值增长了近两倍。 本轮融资由现有投资方菲利斯资本以及加州大学投资部门联合领投。所筹资金将用于两方面:一是为公 ...
8块钱跑通一次强化学习全流程,潞晨云重塑微调赛道:1名算法工程师=1支Infra团队
量子位· 2026-01-07 05:17
Core Viewpoint - The article discusses the shift in large model training from "violent pre-training" to "post-training," emphasizing the importance of fine-tuning and reinforcement learning (RL) in enhancing model performance [1][2]. Group 1: Post-Training and Reinforcement Learning - The industry consensus is that breakthroughs in large model capabilities now rely more on post-training, particularly RL, rather than solely on pre-training parameter accumulation [7]. - DeepSeek-R1's performance improvement in AIME mathematical reasoning benchmark, with pass@1 increasing from 15.6% to 77.9% through RL, exemplifies the potential of RL in achieving significant capability leaps with limited data [7]. Group 2: Challenges in Algorithm Engineering - Algorithm engineers face significant challenges due to complex distributed infrastructure, high GPU rental costs, and intricate architecture tuning, which hinder access to advanced training environments [3][9]. - The introduction of Tinker aims to simplify the training process by providing a standard API, decoupling algorithm design from infrastructure, allowing developers to focus on data and loss function definitions [10]. Group 3: Efficiency and Cost Structure - The Luchenchun Fine-Tuning SDK allows a single algorithm engineer to replace a large infrastructure team, significantly enhancing productivity by simplifying the training process [12][16]. - The SDK's serverless architecture introduces a "pay-per-token" billing model, which charges users only for effective computation tokens used during prefill, sample, and training, eliminating costs associated with idle GPU time [26][29]. Group 4: Practical Applications and User Experience - The SDK supports various use cases, including academic research, startup MVP validation, and industrial applications, enabling users to conduct experiments without the burden of resource management [32][35][37]. - Users can easily train large models using familiar Python syntax, with the SDK providing a seamless experience from installation to execution, thus lowering the barrier to entry for complex model training [39][41]. Group 5: Future of AI Infrastructure - The ultimate goal of AI infrastructure is to achieve "zero cognitive load," where developers only need to describe data and algorithms, while all operational complexities are managed by the system [42]. - As GPU idle costs approach zero and environment setup times decrease, the efficiency of application innovation will be maximized, pushing the limits of computational capabilities [43].
OpenAI前CTO首个创业产品Tinker,这里全量升级开放了,还有羊毛可薅
机器之心· 2026-01-07 05:16
Core Insights - The article discusses the launch of the Luchenyun Fine-tuning SDK, which is based on the Tinker SDK from Thinking Machines Lab, marking a shift from "craft-style" model training to "industrialized fine-tuning" [1][3][26] - The SDK allows developers to focus on algorithm design while abstracting away the complexities of distributed training infrastructure, enabling a more efficient and cost-effective approach to fine-tuning large models [4][6][26] Group 1: Technological Advancements - The introduction of Tinker SDK simplifies the training process by providing standard APIs for various training functions, allowing developers to define data and loss functions without worrying about infrastructure [4][6] - The SDK supports both supervised fine-tuning (SFT) and complex reinforcement learning (RL) pipelines, enabling users to easily construct training flows using atomic functions [8][24] Group 2: Cost Structure and Efficiency - The Luchenyun SDK adopts a serverless architecture with a "pay-per-token" pricing model, which allows users to only pay for effective computation tokens used during prefill, sampling, and training, while other processes are free [14][18] - This new pricing model significantly reduces wasted budget on non-productive time, as users are no longer charged for GPU usage during data loading or debugging [14][18] Group 3: User Experience and Accessibility - The SDK provides a seamless experience for users, allowing them to work in familiar environments like Jupyter Notebook with standard Python syntax, thus enhancing productivity [8][10] - The system includes an intelligent queue that ensures tasks are executed promptly, with no charges during waiting periods, optimizing resource utilization [12] Group 4: Target Users and Applications - The SDK is designed to cater to various user groups, including researchers who can conduct experiments without worrying about infrastructure, and startups that require rapid validation of MVPs [19][20] - In industrial applications, the SDK allows engineers to define loss logic and reinforcement learning reward functions, providing complete control over model training [21] Group 5: Future Outlook - The article emphasizes that post-training is evolving from an academic niche to a mainstream engineering focus, aiming for a "zero cognitive load" experience for developers [26] - The Luchenyun Fine-tuning SDK is now fully open for use, with promotional offers for early adopters, indicating a push for widespread adoption [27][28]
东方港湾黄海平2025年年报与展望:进化的底色!AI应用的算力需求空间巨大 容得下GPU与TPU一起共治天下
Xin Lang Cai Jing· 2026-01-07 02:19
Group 1 - The capital market continues to be influenced by AI bubble theories, but significant advancements in model capabilities have been observed, particularly with Gemini 3, which surpasses ChatGPT in various evaluations, especially in "multimodal interactive" capabilities [3][45] - The AI industry is experiencing a competitive landscape where companies like OpenAI, Meta, and XAI are racing to enhance their models, with OpenAI planning to release GPT 5.3 in early 2026 to regain its leading position [4][46] - The competition has led to a shift in the tech industry, where companies are increasingly undermining each other rather than collaborating, as seen with OpenAI's entry into advertising and e-commerce, and Google's integration of AI into its search engine [5][47] Group 2 - In 2025, AI capabilities have evolved significantly, with reasoning becoming standard across major language models, and the cost of processing tokens decreasing by 50% [9][50] - Long-term memory capabilities have emerged in AI models, allowing them to remember user interactions and improve task execution strategies, which is essential for developing personal assistant applications [10][50] - The concept of "craft intelligence" has developed, where AI is expected to deliver satisfactory results in various tasks, reflecting a shift from merely providing accurate answers to replicating human best practices [11][51] Group 3 - The economic value generated by AI is complex, with significant investments in AI data centers (AIDC) expected to reach nearly $500 billion in 2025, leading to substantial depreciation costs for companies [15][16] - The revenue generated from AI applications is difficult to quantify, as it is spread across cloud vendors and enterprises that utilize AI tokens for internal improvements [17][19] - Companies are increasingly purchasing AI applications rather than building them in-house, with 76% of enterprises opting for external solutions in 2025, indicating a rapid acceptance of AI applications in the market [19][21] Group 4 - The future of AI applications is expected to bring transformative changes, including significant improvements in model performance and the potential for traditional software paradigms to be disrupted [23][25] - The integration of multimodal capabilities in AI models is anticipated to redefine content creation, moving towards an "experience industry" where video and interactive content become prevalent [32][34] - The demand for computational power in AI is projected to grow exponentially, with GPU and TPU technologies competing for dominance in the market [36][38]
北大90后副教授董豪出任上纬启元首席科学家, 研究方向聚焦具身智能
Xin Lang Cai Jing· 2026-01-06 13:28
1月6日,据第一财经,北京大学90后长聘副教授董豪加入上纬新材的个人机器人业务品牌上纬启元,并 出任上纬启元的首席科学家。上纬启元表示,董豪将聚焦具身智能模型领域的技术研发与战略布局。 公开资料显示,董豪现任北京大学计算机学院前沿计算研究中心的长聘副教授,是科技创新2030国家重 大项目负责人(首席科学家),入选国家级高层次青年人才计划。研究方向聚焦于具身智能、大模型、 强化学习、计算机视觉以及相应的开源系统。核心目标是探索并构建兼具成本效益与通用性的机器人系 统。长期以来,始终深耕开源AI系统领域,牵头负责了多个开源项目。 出任上纬启元的首席科学家相关信息,已在董豪个人学术主页中更新确认。 业内分析认为,董豪教授在具身智能领域的深厚积淀,与上纬启元的技术研发需求高度契合,其加入将 进一步强化企业在核心算法、通用机器人系统等关键领域的竞争力。此次引入首席科学家,也表明上纬 新材正在强化其在消费级人形机器人的顶层科研布局。 今年7月8日,上纬新材公告,智元机器人拟通过其与核心团队共同出资设立的持股平台,以"协议转让 +要约收购"组合方式获取公司控制权。11月25日晚,上纬新材公告,公司第四届董事会第一次会议,全 ...
开年收到了很多同学关于自驾方向选择的咨询......
自动驾驶之心· 2026-01-06 09:17
Core Insights - The article emphasizes the importance of deep learning in the fields of automation and computer science, particularly for students in these areas to explore cutting-edge topics such as VLA, end-to-end learning, and world models [2][3] - It highlights the need for newcomers to engage with research papers and discussions to develop their own ideas and methodologies [2] - The article introduces a paper guidance service aimed at assisting students with various aspects of research paper writing and publication [3][4][6] Group 1 - The article suggests that students from computer science and automation backgrounds should focus on deep learning, with specific recommendations for topics like VLA, end-to-end learning, and world models [2] - For mechanical and vehicle engineering students, it recommends starting with traditional PnC and 3DGS due to their lower computational requirements and ease of entry [2] - The article encourages new researchers to learn from failures and emphasizes the importance of developing personal insights through extensive reading and communication [2] Group 2 - The paper guidance service offers support in selecting research topics, full process guidance, and experimental assistance [6] - The service has a high acceptance rate for papers submitted to top conferences and journals, including CVPR, AAAI, and ICLR [7] - Pricing for the guidance service varies based on the level of the paper, and further details can be obtained by contacting the research assistant [8]
海尔消费金融2025年“特征英雄”落下帷幕,数智化风控质效显著
Sou Hu Cai Jing· 2026-01-06 07:50
Core Insights - Haier Consumer Finance successfully concluded its 2025 "Feature Hero" initiative, aimed at enhancing data-driven value in financial services and expanding multi-dimensional data samples [1][6] - The initiative emphasizes the importance of data and features in risk control, with advanced models and algorithms striving to approach the risk identification "ceiling" determined by data [1] Group 1: Feature Hero Competition - The first prize of the "Feature Hero" competition was awarded to the Risk Management Center, which innovatively utilized large models to replace manual processing of voice data, aiding in credit risk control strategies [5] - The competition attracted 32 employees, resulting in the extraction of 2,023 high-quality features from vast data, significantly enhancing the risk control system [5] Group 2: Intelligent Risk Control System - By 2025, Haier Consumer Finance's intelligent risk control system had launched a total of 10,427 real-time features, a 70% increase year-on-year [6] - The company emphasizes the importance of continuous competitions like "Feature Hero" to foster an AI-driven culture and enhance data asset exploration [6] Group 3: AI Integration and Industry Trends - The integration of deep learning technologies such as large models, graph learning, and natural language processing is transforming credit risk control models, showcasing a trend of multi-technology application in the field [6] - Haier Consumer Finance's AI-driven risk control system significantly reduces fraud risk and improves credit approval efficiency, achieving a dual advantage of controllable risk and efficient service [6] Group 4: Future Developments - Future advancements in technologies like federated learning, reinforcement learning, and AGI are expected to further enhance risk control models in areas such as data privacy protection and dynamic strategy optimization [7] - The company plans to deepen its AI First strategy, continuously strengthening data governance and technical application capabilities for high-quality development in credit business [7]