Alternating Training-based Label Smoothing Enhances Prompt Generalization

📅 2025-08-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Prompt tuning in vision-language models suffers from poor generalization, and conventional label smoothing (LS) further degrades performance. To address this, we propose Alternating Training-based Label Smoothing (AT-LS). AT-LS alternates optimization between hard and soft labels to mitigate the adverse impact of standard LS on prompt tuning. It further introduces class-level and instance-level offline soft labels to explicitly model inter-class semantic relationships and intra-class instance discriminability. Theoretical analysis and an efficient soft-label generation strategy ensure seamless compatibility with mainstream prompt-learning frameworks. Extensive experiments demonstrate that AT-LS consistently enhances the generalization of diverse prompt-tuning methods across multiple benchmarks, yielding average improvements of 1.2–2.8 percentage points. AT-LS is plug-and-play, framework-agnostic, and exhibits strong cross-task generalizability.

Technology Category

Application Category

📝 Abstract
Recent advances in pre-trained vision-language models have demonstrated remarkable zero-shot generalization capabilities. To further enhance these models' adaptability to various downstream tasks, prompt tuning has emerged as a parameter-efficient fine-tuning method. However, despite its efficiency, the generalization ability of prompt remains limited. In contrast, label smoothing (LS) has been widely recognized as an effective regularization technique that prevents models from becoming over-confident and improves their generalization. This inspires us to explore the integration of LS with prompt tuning. However, we have observed that the vanilla LS even weakens the generalization ability of prompt tuning. To address this issue, we propose the Alternating Training-based Label Smoothing (ATLaS) method, which alternately trains with standard one-hot labels and soft labels generated by LS to supervise the prompt tuning. Moreover, we introduce two types of efficient offline soft labels, including Class-wise Soft Labels (CSL) and Instance-wise Soft Labels (ISL), to provide inter-class or instance-class relationships for prompt tuning. The theoretical properties of the proposed ATLaS method are analyzed. Extensive experiments demonstrate that the proposed ATLaS method, combined with CSL and ISL, consistently enhances the generalization performance of prompt tuning. Moreover, the proposed ATLaS method exhibits high compatibility with prevalent prompt tuning methods, enabling seamless integration into existing methods.
Problem

Research questions and friction points this paper is trying to address.

Improve prompt tuning generalization in vision-language models
Address label smoothing weakening prompt tuning performance
Enhance adaptability of models to downstream tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Alternating training with one-hot and soft labels
Introducing class-wise and instance-wise soft labels
Seamless integration with existing prompt tuning methods
🔎 Similar Papers
No similar papers found.
Y
Yang Chen
Southern University of Science and Technology, Shenzhen, China
Y
Yanbin Wei
Southern University of Science and Technology, Shenzhen, China; Hong Kong University of Science and Technology, Hong Kong, China
Ke Jin
Ke Jin
Professor at Beijing Institute of Technology
Radiation damageIon Beam Analysishigh entropy alloysNuclear Material
Y
Yi Kong
China University of Mining and Technology, Xuzhou, China
J
James Kwok
Hong Kong University of Science and Technology, Hong Kong, China
Y
Yu Zhang
Southern University of Science and Technology, Shenzhen, China