Leveraging Self-Attention for Input-Dependent Soft Prompting in LLMs
Abstract
A new method using input-dependent soft prompting with a self-attention mechanism improves parameter-efficient fine-tuning for large language models, enhancing zero-shot domain transfer.
The performance of large language models in domain-specific tasks necessitates fine-tuning, which is computationally expensive and technically challenging. This paper focuses on parameter-efficient fine-tuning using soft prompting, a promising approach that adapts pre-trained models to downstream tasks by learning a small set of parameters. We propose a novel Input Dependent Soft Prompting technique with a self-Attention Mechanism (ID-SPAM) that generates soft prompts based on the input tokens and attends different tokens with varying importance. Our method is simple and efficient, keeping the number of trainable parameters small. We show the merits of the proposed approach compared to state-of-the-art techniques on various tasks and show the improved zero shot domain transfer capability.
Community
๐ฏ ID-SPAM (Input-Dependent Soft Prompting technique with a self-Attention Mechanism) is here! ๐
๐ง Efficiently adapt LLMs with input-aware soft prompts using self-attention
โก Minimal parameters, maximum adaptability โ say goodbye to heavy fine-tuning!
๐ Superior zero-shot domain transfer across diverse tasks
๐ Accepted at ACL 2025 (Main) Conference
๐ ID-SPAM learns to generate smarter prompts by attending to input tokens with varying importance, outperforming state-of-the-art parameter-efficient tuning methods. Compact, scalable, and ready for real-world domains!
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Text-to-LoRA: Instant Transformer Adaption (2025)
- MoRE: A Mixture of Low-Rank Experts for Adaptive Multi-Task Learning (2025)
- MAP: Revisiting Weight Decomposition for Low-Rank Adaptation (2025)
- Efficient Knowledge Transfer in Multi-Task Learning through Task-Adaptive Low-Rank Representation (2025)
- Contrastive Prompting Enhances Sentence Embeddings in LLMs through Inference-Time Steering (2025)
- RepCali: High Efficient Fine-tuning Via Representation Calibration in Latent Space for Pre-trained Language Models (2025)
- Adaptive Task Vectors for Large Language Models (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper