主动式上下文工程

Search documents
微调已死?Agentic上下文工程登场,无需微调实现模型进化
机器之心· 2025-10-11 03:29
Core Insights - The article discusses a new technique called Agentic Context Engineering (ACE) that allows language models to self-improve without the need for fine-tuning [1][9]. Context Adaptation - Modern AI systems based on large language models (LLMs) increasingly rely on context adaptation, which enhances model performance by introducing clearer instructions and structured reasoning steps post-training [4]. - Context adaptation offers several advantages over parameter updates, including better interpretability for users and developers, rapid integration of new knowledge, and the ability to share across multiple models or modules [4]. Limitations of Existing Methods - Two main limitations of current context adaptation methods are identified: 1. Brevity bias, where optimization tends to favor concise instructions, potentially overlooking critical domain-specific heuristics [5]. 2. Context collapse, where reliance on LLMs to rewrite prompts leads to degradation into shorter, vaguer summaries over time, negatively impacting performance [6]. Introduction of ACE - ACE is proposed as a solution to these limitations, viewing context as a dynamic, evolving "playbook" rather than a static summary [8][12]. - The framework supports both offline and online scenarios, allowing for scalable and efficient context adaptation [11]. Key Innovations of ACE - ACE introduces three collaborative roles: Generator, Reflector, and Curator, mimicking human learning processes [16]. - The workflow involves the Generator creating reasoning trajectories, the Reflector distilling insights from successes and failures, and the Curator integrating these insights into structured context updates [17]. Incremental Delta Updates - ACE represents context as a collection of structured entries rather than a single prompt, allowing for localized updates and maintaining old knowledge while absorbing new insights [18][20]. - This design leads to reduced computational costs and delays, as ACE generates compact incremental contexts instead of rewriting the entire context [20]. Grow-and-Refine Mechanism - The Grow-and-Refine process ensures that context remains compact and relevant by periodically distilling new entries and updating existing ones [21][22]. - Redundancy is eliminated through semantic embedding comparisons, maintaining the dynamic scalability and high relevance of the context [23][25]. Performance of ACE - Experiments show that ACE significantly outperforms baseline methods in both agent tasks and domain-specific tasks, achieving higher accuracy, faster adaptation, and lower computational costs [29][30]. - In the AppWorld benchmark, ACE improved performance by up to 17.1% without labeled data, bringing open-source models closer to commercial systems [35]. Domain-Specific Task Improvement - In complex financial reasoning tasks, ACE constructed a rich knowledge "playbook," resulting in an average performance increase of 8.6% [40]. Cost and Latency Analysis - ACE demonstrated a significant reduction in adaptation latency by an average of 86.9% and decreased generation costs, showcasing its efficiency [44]. Implications for Continuous Learning - ACE offers a flexible and efficient alternative to traditional model fine-tuning, allowing for context updates that are generally less costly and more interpretable [47]. - The framework is seen as a potential core mechanism for promoting continuous and responsible learning in AI systems [48].