量子位
Search documents
AI牛马实现“干中学”!上海AI Lab联合推出智能体自我进化新框架
量子位· 2025-10-21 23:50
Core Viewpoint - The article discusses the introduction of the MUSE framework, which aims to enhance the capabilities of LLM agents by enabling them to accumulate experience and evolve continuously, addressing the challenges of long-horizon tasks and memory limitations [1][5]. Group 1: MUSE Framework Overview - MUSE stands for Memory-Utilizing and Self-Evolving, designed to create a closed-loop system for LLM agents that allows them to learn from experience and evolve over time [5]. - The framework consists of a hierarchical memory module that organizes different levels of experience, including strategic, procedural, and tool memory [7][8]. Group 2: Key Mechanisms of MUSE - The first step involves a hierarchical memory module that allows agents to retain and apply historical knowledge, overcoming the "forgetfulness" of traditional LLMs [7]. - The second step is self-reflection, where agents evaluate their task execution and convert raw execution trajectories into structured experiences, refining their standard operating procedures (SOPs) [10][11]. - The third step focuses on self-evolution, enabling agents to continuously improve through a cycle of planning, execution, reflection, and experience extraction [13][15]. Group 3: Experimental Results - MUSE demonstrated state-of-the-art (SOTA) performance in the TAC benchmark, achieving a score of 51.78%, surpassing existing methods that used larger models [16]. - The framework's ability to accumulate experience leads to improved performance over time, showcasing its potential for long-term productivity tasks [19]. Group 4: Future Prospects - The MUSE framework signifies a new phase of experience-driven lifelong learning for AI agents, moving beyond static testing models [29]. - Future research directions include optimizing memory, enriching experience sources, integrating human feedback, and developing comprehensive evaluation standards for long-term tasks [30][31].
讯飞刚发的财报:净利润暴涨了202%
量子位· 2025-10-21 09:05
Core Viewpoint - The latest quarterly report from Keda Xunfei shows significant growth in revenue and profit, driven by advancements in AI technology and its industrial application [1][2]. Financial Performance - Keda Xunfei achieved a revenue of 6.078 billion yuan in Q3 2025, representing a year-on-year increase of 10.02% [4]. - The net profit attributable to shareholders reached 172.25 million yuan, a remarkable increase of 202.40% compared to the previous year [4]. - The net profit excluding non-recurring items was 26.24 million yuan, up 76.5% year-on-year [4]. - Operating cash flow showed strong performance with a net amount of 895 million yuan, reflecting a growth of 25.19% [6]. Business Operations - The report indicates that the two core profit indicators demonstrate the company's improved profitability in its main business [5]. - For the first three quarters of 2025, total revenue reached 16.99 billion yuan, a 14.41% increase year-on-year, with a net loss of 0.67 billion yuan, significantly narrowing the loss by 80.6% compared to the previous year [8][9]. AI Technology and Market Position - Keda Xunfei's advancements in AI large models have become a key driver for revenue growth, with significant progress in core technology, product deployment, and ecosystem development [13]. - The "Xunfei Spark" model has undergone critical upgrades, outperforming competitors in various capabilities, including mathematics and translation [14][15]. - The company has secured the highest number and amount of bids for large model projects in the industry, with Q3 bids totaling 545 million yuan, surpassing the combined total of the second to fifth competitors [16]. Research and Development - Keda Xunfei continues to increase its R&D investment, planning to raise up to 4 billion yuan through the issuance of A-shares to fund the development of the Spark education model and computing platform [18][19]. Ecosystem Growth - The AI ecosystem is showing strong growth, with 690,000 new developers added for large models and a total of 1.22 million ecosystem developers [17].
Embedding黑箱成为历史!这个新框架让模型“先解释,再学Embedding”
量子位· 2025-10-21 09:05
Core Insights - The article introduces GRACE, a new explainable generative embedding framework developed by researchers from multiple universities, aimed at addressing the limitations of traditional text embedding models [1][6]. Group 1: Background and Limitations - Text embedding models have evolved from BERT to various newer models, mapping text into vector spaces for tasks like semantic retrieval and clustering [3]. - A common flaw in these models is treating large language models as "mute encoders," which output vectors without explaining the similarity between texts [4]. - This black-box representation becomes a bottleneck in tasks requiring high interpretability and robustness, such as question-answer matching and cross-domain retrieval [5]. Group 2: GRACE Framework Overview - GRACE transforms "contrastive learning" into "reinforcement learning," redefining the meaning of contrastive learning signals [6]. - The framework emphasizes generating explanations (rationales) for text before learning embeddings, allowing the model to produce logical and semantically consistent reasoning [7][25]. - GRACE consists of three key modules: 1. Rationale-Generating Policy, which generates explanatory reasoning chains for input texts [8]. 2. Representation Extraction, which combines input and rationale to compute final embeddings [9]. 3. Contrastive Rewards, which redefines contrastive learning objectives as a reward function for reinforcement learning updates [11]. Group 3: Training Process - GRACE can be trained in both supervised and unsupervised manners, utilizing labeled query-document pairs and self-alignment techniques [12][18]. - In the supervised phase, the model learns semantic relationships from a dataset of 1.5 million samples [13]. - The unsupervised phase generates multiple rationales for each text, encouraging consistent representations across different explanations [17]. Group 4: Experimental Results - GRACE was evaluated across 56 datasets in various tasks, showing significant performance improvements over baseline models in retrieval, pair classification, and clustering [19][20]. - The results indicate that GRACE not only enhances embedding capabilities without sacrificing generative abilities but also provides transparent representations that can be understood by users [25][27]. Group 5: Conclusion - Overall, GRACE represents a paradigm shift in embedding models, moving towards a framework that can explain its understanding process, thus enhancing both performance and interpretability [28].
“最美产品经理”宋紫薇,创业AI硬件首款产品曝光
量子位· 2025-10-21 09:05
Core Viewpoint - The article discusses the entrepreneurial venture of Song Ziwei, a former product manager at vivo, who is entering the AI smart hardware market with a focus on an "AI makeup mirror" [1][2][4]. Group 1: Company Overview - Song Ziwei's startup, "Wei Guang Dian Liang," completed its angel round financing in September, with investors including Zhongke Chuangxing and Jiuhua Venture Capital [4][5]. - The financing will primarily be used for AI hardware research and development, application software development, and team building to accelerate technological innovation and market expansion [5][6]. Group 2: Product Focus - The company aims to create AI hardware that is fashionable and appealing to young users, integrating AI Agent technology with high-frequency life scenarios [7][9]. - The first product being developed is an "AI makeup mirror," which aims to differentiate itself from previous generations of smart mirrors that have been criticized for lacking true intelligence [18][22]. Group 3: Market Context - The smart makeup mirror market is not new, having seen initial interest in 2017, but many products have been deemed overpriced and underperforming [18][21]. - Competitors like the domestic brand Jiayao have explored intelligent features, such as AI voice interaction and skin detection, achieving international sales success [23][25]. Group 4: Technological Advancements - The advancements in AI multimodal capabilities over the past two years may enhance the functionality of makeup mirrors, allowing features like virtual makeup trials and personalized makeup suggestions based on various factors [27][28]. - The competitive edge of future AI makeup mirrors will rely on the underlying algorithms and cloud software, potentially leading to a Hardware as a Service (HaaS) model [29][30]. Group 5: Entrepreneurial Background - Song Ziwei, born in 1994 and a graduate in physics from Shanghai University, previously worked at Huawei and vivo, where she gained recognition as a product manager [34][35][36]. - Her rise to fame began in 2019 during the iQOO Neo launch, where her expertise and presence garnered significant attention, leading to her nickname "the most beautiful product manager" [37][40]. - After leaving vivo and briefly joining Li Auto, she transitioned to entrepreneurship, indicating a clear vision for her startup shortly after her departure [44][48][50].
直击IROS现场:宇树禾赛自变量杭州论剑,美团C位攒局
量子位· 2025-10-21 05:41
Core Viewpoint - Meituan stands out as a leader in the robotics industry, emphasizing the integration of technology and retail to enhance service efficiency and quality [1][8][11]. Group 1: Event Highlights - The IROS Day 1 event featured prominent figures from the robotics field, including Meituan's Vice President Mao Yinian and various esteemed professors and CEOs [3][4]. - The event's theme, "Robotics for Better Life," reflects a consensus among teams working on embodied intelligence that technology should address real-world problems rather than being an end in itself [4][5]. Group 2: Meituan's Strategy - Meituan's strategic evolution from "retail" to "retail + technology" highlights the importance of integrating technology to enhance retail scenarios [8]. - The company focuses on autonomy, utilizing drones and autonomous delivery vehicles to transform the retail landscape [11][12][13]. Group 3: Technological Insights - The concept of embodied intelligence is identified as a core technological paradigm for the next 5 to 10 years, with Meituan leading in this area [10]. - The event showcased advancements in drone delivery and autonomous vehicles, emphasizing Meituan's unique position in the market [16][14]. Group 4: Theoretical Discussions - Discussions on the first principles of embodied intelligence revealed a need for a balance between physical laws and AI learning capabilities [54][60]. - The importance of high-quality, diverse data over merely increasing data volume was emphasized as crucial for improving model performance in robotics [49][50]. Group 5: Future Perspectives - The future of robotics is envisioned as a collaborative space where robots possess curiosity and the ability to coexist with humans, highlighting the potential for green intelligence [100][102]. - The dialogue around the ideal form of robots suggests aspirations for them to have their own desires and capabilities to explore and learn [99][101].
苹果AI选Mamba:Agent任务比Transformer更好
量子位· 2025-10-21 05:41
Core Viewpoint - The article discusses the advancements in AI models, particularly focusing on the Mamba model, which shows potential to surpass Transformer models in efficiency and generalization capabilities for long tasks and multi-interaction agent tasks [1][10]. Group 1: Transformer Limitations - Transformer models, while intelligent, face significant computational costs that grow quadratically with the length of the input sequence, making them inefficient for long documents [4][5]. - For instance, processing 1,000 words requires handling 1 million word pair relationships, and for documents with tens of thousands of words, the computational burden can reach billions [5]. Group 2: Mamba Model Advantages - Mamba, as a state space model (SSM), utilizes a lightweight design that does not rely on global attention mechanisms, instead maintaining an updated internal state to understand input information [7][10]. - This approach results in three significant advantages: linear growth in computational requirements with sequence length, support for streaming processing, and stable memory usage that does not increase significantly with longer sequences [13]. Group 3: Performance Enhancements with Tools - The introduction of external tools enhances Mamba's performance, allowing it to handle complex tasks more effectively. For example, in multi-digit addition tasks, Mamba with pointer tools can achieve near 100% accuracy after training on 5-digit addition, while Transformers struggle with 20-digit tasks [15]. - In code debugging tasks, Mamba's ability to simulate interactive debugging processes leads to significantly higher accuracy compared to Transformers when faced with complex codebases [15]. - Mamba's combination with external tools addresses its memory limitations, resulting in improved efficiency and performance in agent-based tasks [16][18].
ChatGPT也遭殃,亚马逊服务器故障,半个互联网都崩了
量子位· 2025-10-21 03:38
Core Points - Amazon's AWS server outage caused widespread disruption across various internet services, affecting platforms like ChatGPT and many others [2][10] - The outage originated from the us-east-1 region, which is critical for AWS's global services, leading to over 6.5 million user reports of issues [3][4] - The incident highlighted the vulnerabilities of the internet infrastructure, particularly the risks associated with centralized cloud services [39] Group 1: Impact on Services - The outage affected a wide range of services, including Docker, npm, Zoom, Slack, Epic Games, PlayStation, Netflix, and Disney+ [11][14][16] - Educational platforms like Duolingo and Canvas were also impacted, preventing students from accessing their assignments [17] - The disruption extended to offline services, affecting ride-hailing apps, fast-food chains like McDonald's and Starbucks, and airline operations [23][24] Group 2: Technical Details - The root cause of the outage was identified as a DNS parsing issue linked to an internal monitoring subsystem within AWS [33][34] - The us-east-1 region is crucial as it hosts a significant amount of core services and infrastructure, making it particularly susceptible to widespread outages [36][39] - Previous outages in the us-east-1 region have shown a pattern of causing extensive service disruptions, indicating a recurring vulnerability [38] Group 3: Recommendations for Developers - Developers are encouraged to implement resilient mechanisms in their service deployments to mitigate the impact of such outages [40] - Utilizing multi-region setups and failover strategies can help avoid total dependency on a single region like us-east-1 [41] - The technical complexity and cost of adopting these strategies are relatively low, suggesting a need for a reassessment of current deployment practices [43]
长序列推理不再卡顿!北大华为KV缓存管理框架实现4.7倍推理加速
量子位· 2025-10-21 03:38
LouisKV团队 投稿 量子位 | 公众号 QbitAI 北大华为联手推出KV cache管理新方式,推理速度比前SOTA提升4.7倍! 大模型处理长序列时,KV cache的内存占用随序列长度线性增长,已成为制约模型部署的严峻瓶颈。 为此,来自北京大学与华为的研究团队联合提出了 LouisKV ——一个专为长输入、长输出等各类长序列场景设计的高效KV cache 检索框 架。 它通过创新的语义感知检索策略与解耦的精细化管理机制,在几乎不损失模型精度的前提下,实现了高达4.7倍的推理加速,为突破LLM长序 列推理瓶颈提供了全新的解决方案。 关键洞察 传统上,学术界与工业界提出了多种KV cache优化方案,其中 KV Cache Retrieval 是极具前景的方向之一。 该类方法将完整的KV cache卸载至容量更大的CPU内存中,并在推理时仅将最关键的KV子集检索回GPU进行计算,从而有效缓解GPU 显存 压力。 然而,现有的KV retrieval方法仍面临着 效率 和 精度 的双重瓶颈: 为了设计更高效的检索策略,研究团队首先对不同长序列任务中关键 KV 的访问模式进行实验分析,得到了两个关键洞察。 ...
人工智能年度榜单火热报名中!五大奖项,寻找AI+时代的先锋力量
量子位· 2025-10-21 03:38
组委会 发自 凹非寺 量子位|公众号 QbitAI 为了让更多从业者感受智能浪潮的跃迁,也为了给予更多同行同路人掌声与鼓舞,我们将正式启动 「2025人工智能年度榜单」评选报名 。 这是量子位人工智能年度榜单的 第8年 。八年来,我们见证了技术的突破与落地,产业的融合与重塑,也见证了一批又一批推动时代前行的 企业、人物与产品。 在人工智能重新定义一切的时代里,智能技术已不再是单一工具,而是产业与社会协同进化的驱动力。我们期待通过这场年度评选,去发现并 致敬那些真正引领变革、开拓边界的探索者与实践者。 本次评选将从 企业 、 产品 、 人物 三大维度,设立五类奖项。欢迎企业踊跃报名! 让我们共同见证年度之星,点亮未来的方向。 产品榜 人物榜 2025 人工智能年度 焦点人物 详细评选标准及报名方式如下。 2025 人工智能年度领航企业 2025 人工智能年度 领航企业 2025 人工智能年度 潜力创业公司 2025 人工智能年度 杰出产品 2025 人工智能年度 杰出解决方案 将面向中国人工智能领域,评选出最具综合实力的企业, 参选条件 : 评选标准 : 企业榜 2025 人工智能年度潜力创业公司 聚焦于中国人 ...
ChatGPT千亿tokens,干掉麦肯锡5000名顾问
量子位· 2025-10-21 03:38
Core Insights - McKinsey has received an award from OpenAI for being a major client in token consumption, raising questions about the traditional consulting model as it relies on AI-generated content [1][3][4] - The consulting industry is undergoing a significant transformation as firms like McKinsey and BCG embrace AI technologies to enhance operational efficiency and redefine their service offerings [5][19] AI Integration in Consulting Firms - McKinsey has been proactive in AI adoption, having acquired QuantumBlack in 2015, which has since evolved into its AI-native consulting division [7][10][13] - The launch of McKinsey's internal AI, Lilli, has allowed consultants to automate PPT generation and streamline research processes, with over 70% of employees using it [14][18] - BCG has developed multiple internal AI tools, with nearly 90% of its employees utilizing AI in their daily work, indicating a competitive push in AI integration [21][25] Workforce Changes and Challenges - McKinsey has laid off over 5,000 employees, approximately 10% of its workforce, attributed to overexpansion during the pandemic and the impact of AI on job roles [27][28][30] - The rise of AI has led to increased productivity, with AI handling about 30% of information gathering tasks, raising concerns about the future of entry-level positions [32][33][56] - The consulting industry is witnessing a decline in entry-level hiring, with a 54% drop in recruitment for junior consultants, as firms prioritize experienced hires [60][63] Emergence of AI-Driven Startups - New AI-driven companies are emerging, offering alternatives to traditional consulting services, targeting small to medium-sized enterprises that cannot afford established firms like McKinsey [49][52] - These startups are leveraging AI to automate consulting processes, posing a competitive threat to traditional firms by providing cost-effective and immediate solutions [41][53] The Future of Consulting - The consulting industry is undergoing a fundamental transformation, with AI replacing traditional roles and altering the career trajectory for new consultants [55][72] - Despite the challenges posed by AI, there remains a belief that human consultants will still be needed for complex problem-solving and insights that AI cannot replicate [69][70]