🤖 AI Summary
Current foundation models exhibit significant limitations in continual learning, experience accumulation, and personalization, with conventional weight-update-based approaches particularly prone to catastrophic forgetting. This work proposes a novel architecture centered on modular memory that systematically integrates in-weight learning (IWL) and in-context learning (ICL) for the first time: ICL enables rapid acquisition of new knowledge, while IWL provides stable, long-term capability enhancement, and modular memory serves as a synergistic bridge between the two. The resulting framework offers an innovative and practical pathway toward building agents capable of lifelong evolution, continuous adaptation, and personalized interaction.
📝 Abstract
Foundation models have transformed machine learning through large-scale pretraining and increased test-time compute. Despite surpassing human performance in several domains, these models remain fundamentally limited in continuous operation, experience accumulation, and personalization, capabilities that are central to adaptive intelligence. While continual learning research has long targeted these goals, its historical focus on in-weight learning (IWL), i.e., updating a single model's parameters to absorb new knowledge, has rendered catastrophic forgetting a persistent challenge. Our position is that combining the strengths of In-Weight Learning (IWL) and the newly emerged capabilities of In-Context Learning (ICL) through the design of modular memory is the missing piece for continual adaptation at scale. We outline a conceptual framework for modular memory-centric architectures that leverage ICL for rapid adaptation and knowledge accumulation, and IWL for stable updates to model capabilities, charting a practical roadmap toward continually learning agents.