Leveraging Self-Attention for Input-Dependent Soft Prompting in LLMs
Abstract
A new method using input-dependent soft prompting with a self-attention mechanism improves parameter-efficient fine-tuning for large language models, enhancing zero-shot domain transfer.
The performance of large language models in domain-specific tasks necessitates fine-tuning, which is computationally expensive and technically challenging. This paper focuses on parameter-efficient fine-tuning using soft prompting, a promising approach that adapts pre-trained models to downstream tasks by learning a small set of parameters. We propose a novel Input Dependent Soft Prompting technique with a self-Attention Mechanism (ID-SPAM) that generates soft prompts based on the input tokens and attends different tokens with varying importance. Our method is simple and efficient, keeping the number of trainable parameters small. We show the merits of the proposed approach compared to state-of-the-art techniques on various tasks and show the improved zero shot domain transfer capability.
Community
๐ฏ ID-SPAM (Input-Dependent Soft Prompting technique with a self-Attention Mechanism) is here! ๐
๐ง Efficiently adapt LLMs with input-aware soft prompts using self-attention
โก Minimal parameters, maximum adaptability โ say goodbye to heavy fine-tuning!
๐ Superior zero-shot domain transfer across diverse tasks
๐ Accepted at ACL 2025 (Main) Conference
๐ ID-SPAM learns to generate smarter prompts by attending to input tokens with varying importance, outperforming state-of-the-art parameter-efficient tuning methods. Compact, scalable, and ready for real-world domains!
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper