Vector Quantization Prompting for Continual Learning

📅 2024-10-27
🏛️ Neural Information Processing Systems
📈 Citations: 5
Influential: 1
📄 PDF
🤖 AI Summary
Catastrophic forgetting in continual learning and the inability of existing prompt-based methods to jointly optimize prompt selection and task loss hinder effective knowledge retention and transfer. Method: This paper proposes a vector quantization (VQ)-based differentiable discrete prompting mechanism—introducing VQ to prompt learning for the first time—to enable end-to-end differentiable optimization of discrete prompts. By freezing the pretrained backbone and jointly optimizing prompt selection and task objectives, it overcomes the non-differentiability bottleneck of conventional identity-based prompt prediction while preserving abstraction and ensuring training stability. Contribution/Results: Adopting a class-incremental paradigm, the method achieves significant improvements over state-of-the-art approaches across multiple benchmarks, effectively mitigating catastrophic forgetting and enhancing cross-task knowledge transfer and generalization capability.

Technology Category

Application Category

📝 Abstract
Continual learning requires to overcome catastrophic forgetting when training a single model on a sequence of tasks. Recent top-performing approaches are prompt-based methods that utilize a set of learnable parameters (i.e., prompts) to encode task knowledge, from which appropriate ones are selected to guide the fixed pre-trained model in generating features tailored to a certain task. However, existing methods rely on predicting prompt identities for prompt selection, where the identity prediction process cannot be optimized with task loss. This limitation leads to sub-optimal prompt selection and inadequate adaptation of pre-trained features for a specific task. Previous efforts have tried to address this by directly generating prompts from input queries instead of selecting from a set of candidates. However, these prompts are continuous, which lack sufficient abstraction for task knowledge representation, making them less effective for continual learning. To address these challenges, we propose VQ-Prompt, a prompt-based continual learning method that incorporates Vector Quantization (VQ) into end-to-end training of a set of discrete prompts. In this way, VQ-Prompt can optimize the prompt selection process with task loss and meanwhile achieve effective abstraction of task knowledge for continual learning. Extensive experiments show that VQ-Prompt outperforms state-of-the-art continual learning methods across a variety of benchmarks under the challenging class-incremental setting. The code is available at href{https://github.com/jiaolifengmi/VQ-Prompt}{this https URL}.
Problem

Research questions and friction points this paper is trying to address.

Overcoming catastrophic forgetting in continual learning
Optimizing prompt selection with task loss
Enhancing task knowledge abstraction with discrete prompts
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses Vector Quantization for discrete prompts
Optimizes prompt selection with task loss
Enhances task knowledge abstraction effectively
🔎 Similar Papers
No similar papers found.
L
Li Jiao
Communication University of China
Qiuxia Lai
Qiuxia Lai
Associate Professor, Communication University of China
Computer Vision
Y
Yu Li
Harbin Institute of Technology, Shenzhen
Q
Qiang Xu
The Chinese University of Hong Kong