Workflow
搜索智能体
icon
Search documents
告别「一条路走到黑」:通过自我纠错,打造更聪明的Search Agent
机器之心· 2025-11-18 05:08
为了攻克这一难题, 腾讯内容算法中心 联合清华大学, 近期提出 ReSeek 框架,它不是对 RAG 的简单改进,而是对 Search Agent 核心逻辑的一次重塑。 ReSeek 的关键在于引入了动态自我修正机制,允许智能体在执行过程中主动评估每一步行动的有效性。一旦发现路径无效或信息错误,它就能及时回溯并探索新 的可能性,从而避免「一条路走到黑」。 为了同时解决 知识的实时性和推理的复杂性 这两大挑战,搜索智能体(Search Agent)应运而生。它与 RAG 的核心区别在于,Search Agent 能够通过与实时搜索 引擎进行多轮交互来分解并执行复杂任务。这种能力在人物画像构建,偏好搜索等任务中至关重要,因为它能模拟人类专家进行深度、实时的资料挖掘。 但 Search Agent 经常面临着一个棘手的瓶颈: 缺乏过程中的自我纠错能力。 现有的智能体一旦在推理早期因一个模糊的查询而走上错误的路径,就会基于这个错 误结果继续执行,引发连锁式错误(Cascading Errors),最终导致整个任务失败。 论文地址:https://arxiv.org/pdf/2510.00568 开源模型及数据集地址: ...
当Search Agent遇上不靠谱搜索结果,清华团队祭出自动化红队框架SafeSearch
机器之心· 2025-10-16 07:34
Core Insights - The article discusses the vulnerabilities of large language model (LLM)-based search agents, emphasizing that while they can access real-time information, they are susceptible to unreliable web sources, which can lead to the generation of unsafe outputs [2][7][26]. Group 1: Search Agent Vulnerabilities - A real-world case is presented where a developer lost $2,500 due to a search error involving unreliable code from a low-quality GitHub page, highlighting the risks associated with trusting search results [4]. - The research identifies that 4.3% of nearly 9,000 search results from Google were deemed suspicious, indicating a prevalence of low-quality websites in search results [11]. - The study reveals that search agents are not as robust as expected, with a significant percentage of unsafe outputs generated when exposed to unreliable search results [12][26]. Group 2: SafeSearch Framework - The SafeSearch framework is introduced as a method for automated red-teaming to assess the safety of LLM-based search agents, focusing on five types of risks including harmful outputs and misinformation [14][21]. - The framework employs a multi-stage testing process to generate high-quality test cases, ensuring comprehensive coverage of potential risks [16][19]. - SafeSearch aims to enhance transparency in the development of search agents by providing a quantifiable and scalable safety assessment tool [37]. Group 3: Evaluation and Results - The evaluation of various search agent architectures revealed that the impact of unreliable search results varies significantly, with the GPT-4.1-mini model showing a 90.5% susceptibility in a search workflow scenario [26][36]. - Different LLMs exhibit varying levels of resilience against risks, with GPT-5 and GPT-5-mini demonstrating superior robustness compared to others [26][27]. - The study concludes that effective filtering methods can significantly reduce the attack success rate (ASR), although they cannot eliminate risks entirely [36][37]. Group 4: Implications and Future Directions - The findings underscore the importance of systematic evaluation in ensuring the safety of search agents, as they are easily influenced by low-quality web content [37]. - The article suggests that the design of search agent architectures can significantly affect their security, advocating for a balance between performance and safety in future developments [36][37]. - The research team hopes that SafeSearch will become a standardized tool for assessing the safety of search agents, facilitating their evolution in both performance and security [37].
搜索范式革命:纳米AI与谷歌的「超级搜索智能体」共识
36氪· 2025-06-12 11:27
Core Viewpoint - The article discusses the evolution of search engines into "super search" intelligent agents by 2025, emphasizing their transition from traditional keyword-based searches to task-oriented engines that understand user intent and deliver actionable solutions [2][8][16]. Group 1: Evolution of Search Engines - The concept of "super search" is moving from theory to reality, with search engines evolving to possess both intent understanding and task execution capabilities [2][3]. - The AI search 1.0 era involved traditional web page ranking with AI enhancements, while AI search 2.0 transitioned to answer engines focused on delivering direct answers [5][8]. - By 2025, AI search 3.0 will enable a closed-loop system where user intent input leads to automatic execution and result delivery, fundamentally changing how users interact with search engines [8][16]. Group 2: Capabilities of Super Search - Super search must incorporate five key capabilities: task planning, multi-model collaboration, high-dimensional information recognition, multi-modal output, and personalized search experiences [9][10][11][12][13]. - Current AI search engines are still in the early stages of development, with notable examples like Nano AI and Google's AI Mode demonstrating varying degrees of these capabilities [14][18]. Group 3: Market Position and Competition - Nano AI has emerged as a leader in the AI search engine market, achieving significant user engagement and outperforming competitors like Perplexity and traditional search engines [19][21]. - The competition in the search engine space is shifting towards more open agent product designs, with companies like Google leveraging their established technology and brand, while Nano AI focuses on rapid innovation and user-centric product development [33][34].
搜索Agent最新高效推理框架:吞吐量翻3倍、延迟降至1/5,还不牺牲答案质量丨南开& UIUC研究
量子位· 2025-05-29 01:08
Core Insights - The article discusses the efficiency challenges faced by AI-driven search agents, particularly those powered by large language models (LLMs), and introduces a new framework called SearchAgent-X that significantly enhances performance [1][3][32]. Efficiency Bottlenecks - The research identifies two main efficiency bottlenecks in search agents: retrieval accuracy and retrieval latency [4][8]. - Retrieval accuracy is not a straightforward relationship; both low and high precision can negatively impact efficiency. Low precision leads to increased rounds of retrieval, while high precision consumes excessive computational resources [5][6][7]. - Search agents benefit from high recall rate approximate searches, which support reasoning without incurring unnecessary costs [7]. Latency Issues - Search agents are highly sensitive to retrieval latency, where even minor increases can lead to significant end-to-end delays, sometimes up to 83 times [11]. - Improper scheduling and retrieval stalls are identified as primary causes of latency, with data showing that up to 55.9% of tokens may be unnecessarily recomputed due to scheduling issues [13]. SearchAgent-X Framework - SearchAgent-X employs two main acceleration mechanisms: priority-aware scheduling and non-stall retrieval [14][16]. - Priority-aware scheduling dynamically prioritizes concurrent requests to minimize unnecessary waiting and redundant computations [17][18]. - Non-stall retrieval allows for flexible, non-blocking searches, enabling early termination of retrieval when results are deemed sufficient [19][20][22]. Performance Improvements - In practical tests, SearchAgent-X demonstrated a throughput increase of 1.3 to 3.4 times and reduced average latency to 20% to 60% of baseline systems [27]. - The framework maintained generation quality comparable to baseline systems, with slight improvements in accuracy observed in some datasets due to the nature of approximate retrieval [28][29]. Technical Contributions - Each optimization component contributes significantly to overall performance, with priority scheduling reducing end-to-end latency by 35.55% and improving cache hit rates [30]. - Non-stall retrieval further enhances cache hit rates and reduces latency, emphasizing the importance of minimizing waiting times in complex AI systems [31]. Future Outlook - The article concludes that future AI systems will require more frequent interactions with external tools and knowledge bases, highlighting the need to address existing efficiency bottlenecks [32][33]. - It emphasizes the importance of balancing the performance of individual tools within the overall workflow of AI agents to avoid compounding delays and inefficiencies [34].