参数化知识
Search documents
刚刚,腾讯姚顺雨署名首篇论文发布,「下半场」先搞上下文学习
机器之心· 2026-02-03 10:35
Core Insights - The core argument of the article emphasizes that the key bottleneck for models to achieve high-value applications lies in their ability to effectively utilize context [1][5][7]. Group 1: Context Learning Challenges - Recent research indicates that even when context is provided, models may still struggle to solve tasks, highlighting a significant shortfall in their learning capabilities [5][32]. - The article discusses the difference in learning abilities among models, comparing it to individuals with varying talents who learn from the same material [5]. - Current models primarily rely on "parameterized knowledge," which is static and does not adapt to new information from the context [12][34]. Group 2: CL-bench Benchmark - The CL-bench benchmark was developed to assess how well language models can learn new knowledge from context and apply it correctly [16][26]. - It includes 500 complex contexts, 1,899 tasks, and 31,607 validation standards, all designed to require models to learn from the provided context [16][27]. - The benchmark covers four main real-world context learning scenarios: domain knowledge reasoning, rule system application, procedural task execution, and empirical discovery [28][29]. Group 3: Model Performance Evaluation - Evaluation results show that even the best-performing model, GPT-5.1 (High), only solved 23.7% of tasks, indicating a significant gap in context learning capabilities [31][32]. - The majority of errors stem from models ignoring or misusing context, rather than a lack of information [34][35]. - The article notes that models struggle particularly with tasks requiring inductive reasoning from experimental data, often achieving less than 10% success [39]. Group 4: Future Directions - The article suggests that improving context learning could shift the role of humans from data providers to context providers in AI systems [43]. - It raises the challenge of how to make knowledge learned from context persistent, as current models lose this knowledge once the context window is cleared [43][46]. - The potential for models to achieve autonomous learning through effective context learning and memory consolidation is highlighted as an exciting future prospect [47][48].