Workflow
上下文自适应
icon
Search documents
微调已死?Agentic上下文工程登场,无需微调实现模型进化
3 6 Ke· 2025-10-11 03:53
Core Insights - A recent paper from Stanford University, SambaNova, and UC Berkeley introduces a technique called Agentic Context Engineering (ACE), which allows language models to self-improve without fine-tuning [1][2]. Context Adaptation - Context adaptation is increasingly relied upon by contemporary AI systems based on large language models (LLMs), enhancing model performance by introducing clearer instructions and structured reasoning steps post-training [2][3]. - Context serves as the foundation for various AI system components, including guiding downstream tasks, memory mechanisms, and factual evidence to reduce hallucinations and supplement knowledge [2]. Advantages of Context Adaptation - Context adaptation offers several core advantages over parameter updates: it is more interpretable for users and developers, allows for rapid integration of new knowledge, and can be shared across multiple models or modules in composite systems [3]. - The advancements in long-context language models and efficient reasoning mechanisms make context-based methods increasingly feasible for building high-performance, scalable, and self-improving AI systems [3]. Limitations of Existing Methods - Existing context adaptation methods face two main limitations: brevity bias, where optimization tends to favor concise instructions at the expense of critical domain-specific heuristics, and context collapse, where reliance on LLMs leads to performance degradation over time [4][6]. Agentic Context Engineering (ACE) - ACE is proposed as a scalable and efficient context adaptation framework that treats context as a dynamic playbook, continuously accumulating and organizing strategies rather than distilling knowledge into static summaries [6][7]. - The ACE framework mimics human learning through an "experiment-reflect-integrate" approach, avoiding bottlenecks associated with single models handling all functions [9]. Key Innovations of ACE - ACE introduces three key roles: Generator (to create reasoning trajectories), Reflector (to distill insights from successes and failures), and Curator (to integrate these insights into structured context updates) [10][11]. - The incremental Delta update mechanism allows for localized updates rather than complete rewrites, significantly reducing latency and computational overhead [11][13]. Performance and Efficiency - ACE has demonstrated superior performance in agent tasks and domain-specific tasks, achieving higher accuracy, faster adaptation, and lower computational costs compared to baseline methods [15][16]. - In experiments, ACE improved performance by up to 17.1% in the AppWorld benchmark and 8.6% in complex financial reasoning tasks [20][25]. Cost and Latency Reduction - ACE significantly reduces adaptation latency by an average of 86.9% and lowers generation costs, making it a more efficient alternative to traditional fine-tuning methods [29][31]. - The framework's design ensures that longer contexts do not lead to increased serving costs, as modern infrastructure optimizes for long-context loads [31]. Future Implications - ACE is positioned as a flexible and efficient alternative to traditional model fine-tuning, potentially becoming a core mechanism for promoting continuous and responsible learning in AI systems [32].
微调已死?Agentic上下文工程登场,无需微调实现模型进化
机器之心· 2025-10-11 03:29
Core Insights - The article discusses a new technique called Agentic Context Engineering (ACE) that allows language models to self-improve without the need for fine-tuning [1][9]. Context Adaptation - Modern AI systems based on large language models (LLMs) increasingly rely on context adaptation, which enhances model performance by introducing clearer instructions and structured reasoning steps post-training [4]. - Context adaptation offers several advantages over parameter updates, including better interpretability for users and developers, rapid integration of new knowledge, and the ability to share across multiple models or modules [4]. Limitations of Existing Methods - Two main limitations of current context adaptation methods are identified: 1. Brevity bias, where optimization tends to favor concise instructions, potentially overlooking critical domain-specific heuristics [5]. 2. Context collapse, where reliance on LLMs to rewrite prompts leads to degradation into shorter, vaguer summaries over time, negatively impacting performance [6]. Introduction of ACE - ACE is proposed as a solution to these limitations, viewing context as a dynamic, evolving "playbook" rather than a static summary [8][12]. - The framework supports both offline and online scenarios, allowing for scalable and efficient context adaptation [11]. Key Innovations of ACE - ACE introduces three collaborative roles: Generator, Reflector, and Curator, mimicking human learning processes [16]. - The workflow involves the Generator creating reasoning trajectories, the Reflector distilling insights from successes and failures, and the Curator integrating these insights into structured context updates [17]. Incremental Delta Updates - ACE represents context as a collection of structured entries rather than a single prompt, allowing for localized updates and maintaining old knowledge while absorbing new insights [18][20]. - This design leads to reduced computational costs and delays, as ACE generates compact incremental contexts instead of rewriting the entire context [20]. Grow-and-Refine Mechanism - The Grow-and-Refine process ensures that context remains compact and relevant by periodically distilling new entries and updating existing ones [21][22]. - Redundancy is eliminated through semantic embedding comparisons, maintaining the dynamic scalability and high relevance of the context [23][25]. Performance of ACE - Experiments show that ACE significantly outperforms baseline methods in both agent tasks and domain-specific tasks, achieving higher accuracy, faster adaptation, and lower computational costs [29][30]. - In the AppWorld benchmark, ACE improved performance by up to 17.1% without labeled data, bringing open-source models closer to commercial systems [35]. Domain-Specific Task Improvement - In complex financial reasoning tasks, ACE constructed a rich knowledge "playbook," resulting in an average performance increase of 8.6% [40]. Cost and Latency Analysis - ACE demonstrated a significant reduction in adaptation latency by an average of 86.9% and decreased generation costs, showcasing its efficiency [44]. Implications for Continuous Learning - ACE offers a flexible and efficient alternative to traditional model fine-tuning, allowing for context updates that are generally less costly and more interpretable [47]. - The framework is seen as a potential core mechanism for promoting continuous and responsible learning in AI systems [48].