Leveraging Self-Attention for Input-Dependent Soft Prompting in LLMs

📅 2025-06-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high computational cost and technical barriers associated with fine-tuning large language models (LLMs), this paper proposes an input-dependent soft prompt method. Our approach leverages self-attention to dynamically weight input tokens and generate learnable soft prompt embeddings conditioned on the current input’s semantic content, enabling lightweight, conditional parameter adaptation. Unlike static or task-level fixed prompts, ours is the first to formulate prompt generation as an input-conditioned process. With negligible parameter overhead (<0.1%), it significantly improves zero-shot cross-domain transfer performance. Extensive experiments across multiple downstream tasks demonstrate consistent superiority over mainstream soft-prompting baselines—including Prefix-Tuning and Prompt-Tuning—validating the synergistic gains from dynamically aligning prompt relevance with input semantics and enhancing generalization.

Technology Category

Application Category

📝 Abstract
The performance of large language models in domain-specific tasks necessitates fine-tuning, which is computationally expensive and technically challenging. This paper focuses on parameter-efficient fine-tuning using soft prompting, a promising approach that adapts pre-trained models to downstream tasks by learning a small set of parameters. We propose a novel Input Dependent Soft Prompting technique with a self-Attention Mechanism (ID-SPAM) that generates soft prompts based on the input tokens and attends different tokens with varying importance. Our method is simple and efficient, keeping the number of trainable parameters small. We show the merits of the proposed approach compared to state-of-the-art techniques on various tasks and show the improved zero shot domain transfer capability.
Problem

Research questions and friction points this paper is trying to address.

Efficient fine-tuning for domain-specific LLM tasks
Input-dependent soft prompting with self-attention
Improving zero-shot domain transfer capability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Input-dependent soft prompting with self-attention
Efficient fine-tuning with minimal parameters
Improved zero-shot domain transfer capability
🔎 Similar Papers
No similar papers found.