Abstract
Soft prompt tuning techniques have recently gained traction as an effective strategy for the parameter-efficient tuning of pretrained language models, particularly minimizing the required adjustment of model parameters. Despite their growing use, achieving optimal tuning with soft prompts, especially for smaller datasets, remains a substantial challenge. This study makes two contributions in this domain: (i) we introduce SUPERPOS-PROMPT, a new reparameterization technique employing the superposition of multiple pretrained vocabulary embeddings to improve the learning of soft prompts. Our experiments across several GLUE and SuperGLUE benchmarks consistently highlight SUPERPOS-PROMPT's superiority over Residual Prompt tuning, exhibiting an average score increase of +6.4 in T5-Small and +5.0 in T5-Base along with a faster convergence. Remarkably, SUPERPOS-PROMPT occasionally outperforms even full fine-tuning methods. (ii) Additionally, we demonstrate enhanced performance and rapid convergence by omitting dropouts from the frozen network, yielding consistent improvements across various scenarios and tuning methods.
| Original language | English |
|---|---|
| Pages (from-to) | 34-46 |
| Number of pages | 13 |
| Journal | Proceedings of Machine Learning Research |
| Volume | 262 |
| DOIs | |
| Publication status | Published - 7 Jun 2024 |
| Event | 4th NeurIPS Efficient Natural Language and Speech Processing Workshop - Vancouver, Canada Duration: 14 Dec 2024 → … |
Fingerprint
Dive into the research topics of 'SuperPos-Prompt: Enhancing Soft Prompt Tuning of Language Models with Superposition of Multi Token Embeddings'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver