Workflow
大语言模型
icon
Search documents
早报|刘强东:近期又约过王兴见面;校方回应男留学生与女生混住;“车顶维权”女车主首赢特斯拉;太二回应门店活鱼现杀争议
虎嗅APP· 2025-09-17 00:20
Group 1 - Microsoft plans to invest over $30 billion in the UK over the next four years, with an additional $15.5 billion for capital expansion on top of the previously announced $3.2 billion for data center infrastructure [2][3] - The investment will also include $15.1 billion for various business initiatives in the UK, such as an AI lab in London and gaming projects [4] Group 2 - A collision involving two XPeng Heitech eVTOLs occurred at the Changchun Airshow, resulting in at least one passenger injury, with the company currently verifying the details [5][6] Group 3 - JD.com's founder Liu Qiangdong announced plans for a new hotel development strategy, emphasizing the need to avoid price wars that could harm service quality and profitability [7][19] Group 4 - The U.S. National Highway Traffic Safety Administration is investigating approximately 174,000 Tesla Model Y vehicles due to potential door handle malfunctions linked to low battery power [8] - A Beijing court ruled that Tesla must provide complete driving data from the 30 minutes prior to an accident, affirming consumer rights to information [15][16] Group 5 - Google's Gemini AI model has surpassed OpenAI's ChatGPT in the Apple App Store's free app rankings, indicating a significant user adoption and interest in Google's advancements in generative AI [9][10] Group 6 - Anta Group reported the dismissal of 74 employees for serious misconduct as part of its anti-corruption efforts, with 46 individuals referred to judicial authorities for criminal activities [13][14] - The company aims to enhance its internal auditing and oversight mechanisms by 2025 [14] Group 7 - Kering Group confirmed a data breach affecting millions of customers from brands like Gucci and Balenciaga, although financial information was not compromised [20] - The company has notified affected customers and reported the incident to relevant data protection authorities [20] Group 8 - The Global Public Safety Cooperation Forum will open in Lianyungang, with over 800 guests from 176 countries and international organizations attending, aiming to release a comprehensive public safety index report [29] - Meta's Connect 2025 conference will focus on the integration of AI glasses and the metaverse, with the anticipated launch of a new smart eyewear product [30] Group 9 - The Chinese government plans to introduce consumption stimulus measures, including appliance trade-in programs and tourism subsidies, to stimulate a trillion-dollar market [31] Group 10 - Ant Group's CEO predicts that large language models may eliminate the need for traditional apps, as intelligent agents take over various tasks, reshaping human-computer interaction [32][33]
起售价23.59万元,奥迪E5 Sportback上市
Bei Jing Shang Bao· 2025-09-16 14:26
Group 1 - Audi has officially launched its first strategic model, the Audi E5 Sportback, with a price range of 235,900 to 319,900 yuan, offering four configurations: Pioneer, Pioneer Plus, Pioneer Quattro, and Flagship Quattro [1] - The Audi E5 Sportback features the new AUDI OS operating system and integrates the Qualcomm Snapdragon 8295 digital cockpit chip, creating an ultra-interactive smart cockpit [3] - The vehicle includes the Audi Assistant, powered by a customized language model "Doubao," which enhances semantic understanding, multi-turn dialogue, and vehicle control interaction capabilities [3] Group 2 - Audi has partnered with Momenta to develop a comprehensive driving assistance solution, covering urban, highway, and parking scenarios [3] - The company plans to accelerate channel development, aiming to establish over 240 fully functional user centers across more than 100 cities in China by the end of this year [3]
IPO研究 | 中国保险AI科技总可触及市场规模预计2029年将达1.35万亿元
Sou Hu Cai Jing· 2025-09-16 10:32
Core Insights - Warmwa Insight Technology Co., Ltd. has submitted a listing application to the Hong Kong Stock Exchange, with JPMorgan and HSBC as joint sponsors [1] - The company is recognized as a leading AI technology firm in China's insurance industry, focusing on underwriting processes and claims management [1] - The insurance industry is undergoing significant transformation driven by technological advancements and data integration, with AI enhancing operational efficiency across the insurance value chain [1] Market Overview - The total addressable market for AI technology in China's insurance sector is projected to reach RMB 746.8 billion in 2024, with an expected growth to RMB 1,353.8 billion by 2029, reflecting a compound annual growth rate (CAGR) of 12.6% from 2024 to 2029 [2] - China's insurance market has been rapidly growing, with premiums increasing from RMB 4.5 trillion in 2020 to RMB 5.7 trillion in 2024, representing a CAGR of 5.9% [3] - The health insurance market is also expanding, with premiums expected to rise from RMB 0.8 trillion in 2020 to RMB 1.0 trillion in 2024, and projected to reach RMB 1.7 trillion by 2029, with a CAGR of 11.6% from 2024 to 2029 [3] Growth Potential - Despite being the second-largest insurance market globally by premium volume in 2023, China's insurance penetration rate stands at only 3.9%, significantly lower than the global average of 7.0% [3] - The insurance density in China is $516, compared to the global average of $889, indicating substantial growth potential in the sector [3]
只要科学任务能打分,AI就能实现SOTA结果 | 谷歌最新论文
量子位· 2025-09-15 05:57
Core Viewpoint - The article discusses a new AI system developed by Google that assists scientists in creating expert-level empirical software, achieving state-of-the-art (SOTA) results across various scientific fields [10][12][30]. Group 1: AI System Development - The AI system utilizes a combination of Large Language Models (LLMs) and tree search algorithms to systematically improve software quality metrics [10][17]. - It addresses the slow and labor-intensive process of developing empirical software, which often takes years to complete [14][15]. - The system can automatically create empirical software for quantifiable tasks, significantly enhancing the efficiency of scientific research [17][24]. Group 2: Performance and Achievements - In bioinformatics, the system discovered 40 novel methods for single-cell data analysis, outperforming top human-developed methods on public leaderboards [25][30]. - In epidemiology, it generated 14 models that surpassed the CDC ensemble and all other individual models for forecasting COVID-19 hospitalizations [10][30]. - The system also produced state-of-the-art software for geospatial analysis, neural activity prediction in zebrafish, time series forecasting, and numerical solutions of integrals [10][30]. Group 3: Methodology and Innovation - The AI system enhances code mutation capabilities by injecting research ideas from highly cited papers, textbooks, and search engine results [21][24]. - It generates numerous candidate software solutions and employs tree search algorithms to filter and optimize these candidates [17][24]. - The integration of complex research ideas allows the system to explore a vast solution space, leading to the discovery of high-quality solutions [24][30]. Group 4: Community Response and Implications - The article notes that the introduction of AI in scientific research has sparked discussions about the appropriateness of delegating research authority to AI [32]. - There are concerns regarding the reliability of AI-generated results and the need for human oversight in the verification process [32][40].
没有专业背景,但他搞出了一家7亿美元估值的公司
Hu Xiu· 2025-09-15 04:49
Core Insights - Legora is rapidly growing in the legal tech sector, having expanded from Europe to the US and partnered with 250 law firms, including top firms like Cleary Gottlieb and Goodwin [1][2] - The company recently raised $80 million in Series B funding, achieving a valuation of $675 million, positioning itself as a strong competitor to Harvey [2] - The founder, Max Junestrand, emphasizes the importance of humility and collaboration with early partners to navigate the rapidly changing legal industry [3] Product Overview - Legora's product consists of a web application and a Word plugin, integrating AI functionalities into Microsoft Word [4] - The web application has evolved from a simple chat feature to a sophisticated intelligent agent capable of managing complex workflows [5][6] - The "Tabular Review" feature allows users to input multiple documents and queries for simultaneous processing, addressing the complexities of legal documents [9][10] Sales Strategy - Legora adopts a "win-win" approach in sales, positioning itself as a long-term partner for law firms needing to adopt new technologies [18][20] - The company recognizes that many legal services are similar, leading to price pressures and a need for efficiency, which drives firms to adopt new technologies [21][22] - Law firms are motivated to become leaders in adopting technology to maintain their competitive edge [23][24] Competitive Landscape - Legora competes with established legal tech companies but believes that the rapid pace of AI development allows it to outpace larger firms in product delivery [41][44] - The company has successfully built a team of around 100 employees, significantly increasing its development speed compared to larger competitors [45][46] - Law firms are increasingly reluctant to commit to long-term contracts, preferring shorter agreements that allow for flexibility in technology adoption [46][47] Future Outlook - The role of lawyers is expected to shift towards being reviewers rather than executors, managing AI outputs and ensuring quality [51][52] - The company aims to be a strategic partner for law firms, helping them navigate the transformation brought about by AI [61] - Junestrand advises new entrants in the legal tech space to avoid being tied to single suppliers and to find unique niches that AI cannot easily penetrate [63][64] Recruitment and Culture - Legora prioritizes hiring individuals with entrepreneurial backgrounds, fostering a culture of creativity and problem-solving [70][72] - The company has expanded from 10 to 100 employees in a year, emphasizing the importance of hiring proactive team members who can leverage AI for greater efficiency [67][68]
告别ROS的繁琐, 易用易学的机器人学习系统: 华为诺亚面向机器人学习的开源Python框架
机器之心· 2025-09-15 04:00
图 1: Ark 的整体框架 近年来,机器人技术在硬件领域取得了显著突破 —— 无论是 DARPA 机器人挑战赛,还是首届人形机器人自由搏击表演,都展示了令人瞩目的进展。然而,机器 人的自主能力仍明显落后于机器学习的发展步伐。 造成这一差距的 关键瓶 颈在于软 件层面 :现有的机器人技术栈学习门槛较高,仍大量依赖 C/C++ 进行底层开发,工具链分散且硬件集成复杂。相比之下,推动 现代人工智能发展的生态系统以 Python 为核心,文档完善、易于使用 —— 两者形成了鲜明对比。 为应对这些挑战,来自 华为诺亚方舟实验室,德国达姆施塔特工业大学,英国伦敦大学学院,帝国理工学院和牛津大学的研究者 们联合推出了 Ark —— 一个基 于 Python 的机器人开 发框架,支持快速原型 构建,并可便捷地在仿真和真实机器人系统上部署新算法 。 Ark 与主流机器学习工作流深度兼容,能够从仿真环境或实际机器人中采集和预处理数据,并支持使用如 ACT、Diffusion Policy 等前沿模仿学习方法进行策略训 练。该框架采用类似 OpenAI Gym 风格的主接口设计,极大降低了机器学习研究者的上手门槛,便于集成与实验 ...
作为研究,VLA至少提供了一种摆脱无尽corner case的可能性!
自动驾驶之心· 2025-09-15 03:56
Core Viewpoint - VLA (Vision-Language-Action) is emerging as a mainstream keyword in autonomous driving, with new players rapidly entering the field and industrial production accelerating, while academia continues to innovate and compete [1][2]. Summary by Sections 1. VLA Research and Development - The VLA model represents a shift from traditional modular architectures to a unified end-to-end model that directly maps raw sensor inputs to driving control commands, addressing previous bottlenecks in autonomous driving technology [3][4]. - Traditional modular architectures (L2-L4) have clear advantages in terms of logic and independent debugging but suffer from cumulative error effects and information loss, making them less effective in complex traffic scenarios [4][5]. 2. VLA Model Advantages - The introduction of VLA models leverages the strengths of large language models (LLMs) to enhance interpretability, reliability, and the ability to generalize to unseen scenarios, thus overcoming limitations of earlier models [5][6]. - VLA models can explain their decision-making processes in natural language, improving transparency and trust in autonomous systems [5][6]. 3. Course Objectives and Structure - The course aims to provide a systematic understanding of VLA, helping participants develop practical skills in model design and research paper writing, while also addressing common challenges faced by newcomers in the field [6][7]. - The curriculum includes 12 weeks of online group research, followed by 2 weeks of paper guidance and 10 weeks of paper maintenance, focusing on both theoretical knowledge and practical coding skills [7][8]. 4. Enrollment and Requirements - The program is designed for a small group of 6 to 8 participants, targeting individuals with a foundational understanding of deep learning and basic programming skills [11][16]. - Participants are expected to engage actively in discussions and complete assignments on time, maintaining academic integrity throughout the course [20][29]. 5. Course Highlights - The course offers a comprehensive learning experience with a multi-faceted teaching approach, including guidance from experienced mentors and a structured evaluation system to track progress [23][24]. - Participants will gain access to essential resources, including datasets and baseline codes, to facilitate their research and experimentation [24][25].
将KV Cache预算降至1.5%!他们用进化算法把大模型内存占用砍下来了
机器之心· 2025-09-14 05:16
Core Insights - EvolKV achieves superior performance with only 1.5% of the full KV cache budget, significantly reducing inference costs for large language models [1][11][25] - The traditional KV cache methods face challenges with long input texts, leading to increased storage requirements and slower processing [3][4] KV Cache Optimization - Existing KV cache compression methods primarily rely on heuristic approaches, which may not optimally retain task-relevant information [4][9] - EvolKV introduces an evolutionary framework that adaptively allocates KV cache budgets across transformer layers, optimizing for downstream task performance [6][10] Performance Improvements - In various benchmark tests, EvolKV consistently outperforms baseline methods, achieving up to a 13% improvement in the Needle-in-a-Haystack benchmark and maintaining high accuracy in the GSM8K dataset [11][30][25] - The method demonstrates strong adaptability across diverse tasks, maintaining competitive performance even with reduced cache budgets [25][29] Experimental Results - Comprehensive experiments on Mistral 7B-Instruct and Llama-3-8B-Instruct show that EvolKV outperforms all baseline methods across multiple KV cache budget configurations [22][24] - In the LongBench evaluation, EvolKV consistently achieved the highest average performance, even surpassing the full model in certain configurations [22][25] Evolutionary Algorithm Mechanism - The evolutionary algorithm generates candidate solutions and evaluates their fitness based on downstream task performance, guiding the optimization process [13][14] - The optimization process is structured in groups to enhance efficiency, allowing for a more stable optimization dynamic [16][17] Cache Budget Allocation - EvolKV employs a dynamic, task-driven approach to allocate KV cache budgets, ensuring that the distribution aligns with the functional contributions of different transformer layers [10][19] - The method includes a mechanism for adjusting the total KV cache budget to ensure fairness in evaluation [20]
AI解数学题只靠最后一个token
量子位· 2025-09-14 05:05
Core Insights - The research indicates that in mental arithmetic tasks, the majority of calculations are concentrated on the last token, rather than being distributed across all tokens, suggesting that global information access is not necessary for specific tasks like mental arithmetic [1][11]. Group 1: Research Methodology - Researchers employed Context-Aware Mean Ablation (CAMA) and attention-based peeking techniques to conduct a series of ablation experiments on models like Llama-3-8B [2][22]. - The experiments aimed to identify the "minimum computation" required for models to perform well by systematically removing or altering parts of the model [3]. - A sparse subgraph termed "All-for-One" (AF1) was identified, which allows efficient computation with minimal layers and limited information transfer [4][5]. Group 2: Model Structure and Functionality - In the AF1 structure, initial layers (L_wait) do not perform calculations related to their own values but instead focus on general preparatory tasks [7]. - Information is transferred to the last token through intermediate layers (L_transfer), which then independently performs the final calculations [8][9]. - This separation of general computation and input-specific computation highlights the model's efficiency in handling arithmetic tasks [10]. Group 3: Experimental Findings - The experiments revealed that Llama-3-8B requires only the first 14 layers for general computation, followed by 2 layers for information transfer, with the remaining layers dedicated to the last token's self-computation [24][26]. - AF1_llama demonstrated high fidelity across eight tasks, maintaining performance levels close to the original model [28][29]. - The importance of specific attention heads in arithmetic calculations was confirmed, with the model retaining approximately 95% accuracy even after removing nearly 60 heads, indicating redundancy in attention heads [30]. Group 4: Generalization and Limitations - AF1_llama was tested for its ability to generalize to other arithmetic forms, showing high accuracy in direct arithmetic tasks but failing in tasks requiring semantic understanding, such as word problems and Python code [32][34]. - Similar AF1-like subgraphs were found in Pythia and GPT-J models, although these models exhibited shorter waiting periods and less clear performance boundaries compared to Llama [35][36]. Group 5: Contributions and Innovations - This research contributes to the understanding of arithmetic reasoning and cross-token computation mechanisms in large language models [37]. - The methodologies introduced, CAMA and ABP, offer innovative approaches that could extend beyond arithmetic tasks to broader applications [37].
Meta开源MobileLLM-R1模型,不到1B参数,用1/10的训练就超越了Qwen3
机器之心· 2025-09-13 08:54
Core Viewpoint - Meta AI has officially released the MobileLLM-R1 series, which includes efficient sub-billion parameter language models optimized for on-device use cases, demonstrating significant performance improvements compared to existing open-source models [4][8]. Group 1: Model Performance and Features - The MobileLLM-R1 series includes three base models: MobileLLM-R1-140M, MobileLLM-R1-360M, and MobileLLM-R1-950M, which are not general chat models but are supervised fine-tuned (SFT) for specific tasks such as mathematics, programming (Python, C++), and scientific questions [6][8]. - The largest model, MobileLLM-R1-950M, was pre-trained using approximately 2 trillion high-quality tokens, achieving performance comparable to models trained on 36 trillion tokens, such as Qwen3 0.6B [8]. - MobileLLM-R1-950M outperforms existing models in various benchmarks, achieving five times higher accuracy on the MATH benchmark compared to the Olmo 1.24B model and twice as high as the SmolLM2 1.7B model [10]. Group 2: Model Architecture and Efficiency - The architecture of the MobileLLM-R1 models includes varying layers and parameters, with MobileLLM-R1-950M having 22 layers and 949 million parameters, while the smaller models have 15 layers and 140 million to 360 million parameters [14]. - The models are designed for text input and output, with a context length of 4k for base models and 32k for final models, supporting a vocabulary size of 128k [15]. Group 3: Research and Development Team - The development of the MobileLLM-R1 series was led by a team of researchers, including Zechun Liu, Ernie Chang, and Changsheng Zhao, who have extensive backgrounds in natural language processing and model optimization [18][21][30]. - The project took a year to develop, focusing on efficient deployment and optimization of large language models for resource-constrained environments [18][22].