嵌套学习
Search documents
LLM 语境下,「持续学习」是否是 「记忆」 问题的最优解?
机器之心· 2025-11-16 01:30
Group 1 - The article discusses the concept of "Nested Learning" proposed by Google, which aims to address the memory management issues in LLMs (Large Language Models) and the challenges of catastrophic forgetting [5][6][8] - Nested Learning is presented as a multi-layered optimization problem, where models are seen as a series of interconnected sub-problems, allowing for the simultaneous learning of new skills while avoiding the loss of previously acquired knowledge [6][7] - The research introduces the "Continuous Memory System" (CMS), which treats memory as a system of multiple modules that update at different frequencies, enhancing the model's ability to manage memory effectively [6][7] Group 2 - The article highlights the importance of improving LLMs' memory capabilities to enable continual learning, allowing AI to retain contextual experiences, semantic knowledge, and procedural skills [8] - A proposed three-layer memory architecture includes Model Weights for general knowledge, KV Cache for intermediate results, and Context for relevant background information, facilitating appropriate responses from the model [8]
Jeff Dean盛赞姚班校友AI新研究,目前人已到Meta
量子位· 2025-11-15 05:00
Core Viewpoint - The article discusses a new paradigm in AI called Nested Learning (NL), which addresses the issue of catastrophic forgetting in large language models and proposes a more efficient learning structure that mimics human cognitive processes [2][10][25]. Summary by Sections Nested Learning Concept - Nested Learning transforms models from a flat computational network to a hierarchical, self-adjusting learning system, inspired by the human brain's memory processes [6][12][14]. - Traditional models like Transformers are seen as simplified versions of NL, lacking the multi-level advantages that NL offers [6][14]. Innovations of Nested Learning - The research team introduced three core innovations based on NL: 1. **Deep Optimizer**: Unlike traditional optimizers, NL's deep optimizer uses a pre-processing mechanism to understand gradient properties and employs MLP neural networks for memory, allowing for flexible parameter adjustments [17][18]. 2. **Self-Modifying Model**: This allows models to autonomously learn how to adjust their parameters during training, adapting to new data without manual intervention [19]. 3. **Continuous Memory System**: Upgrades the traditional short-term/long-term memory structure to a multi-scale memory chain, enabling efficient storage and processing of information [20]. Performance of Hope Model - The Hope model, based on NL, significantly outperforms mainstream baseline models like Transformer, RetNet, and DeltaNet in language modeling and common-sense reasoning tasks, demonstrating lower perplexity and higher accuracy across various metrics [8][23][24]. - For instance, in language modeling tasks, Hope achieved a perplexity of 26.05 with 760M parameters, outperforming other models [24]. Implications of Nested Learning - The introduction of NL represents a paradigm shift in deep learning, moving away from the traditional approach of stacking layers and parameters, and instead leveraging cognitive science to create a collaborative, hierarchical intelligence system [25]. - This new paradigm may enable AI to continuously learn and accumulate knowledge like humans, potentially solving key challenges in long-context reasoning and lifelong learning [25].