SuperPos-Prompt: Enhancing Soft Prompt Tuning of Language Models with Superposition of Multi Token Embeddings

Mohammad Ali SadraeiJavaeri*, Ehsaneddin Asgari, Alice Carolyn McHardy, Hamid Reza Rabiee

*Corresponding author for this work

Research output: Contribution to journalConference articlepeer-review

Abstract

Soft prompt tuning techniques have recently gained traction as an effective strategy for the parameter-efficient tuning of pretrained language models, particularly minimizing the required adjustment of model parameters. Despite their growing use, achieving optimal tuning with soft prompts, especially for smaller datasets, remains a substantial challenge. This study makes two contributions in this domain: (i) we introduce SUPERPOS-PROMPT, a new reparameterization technique employing the superposition of multiple pretrained vocabulary embeddings to improve the learning of soft prompts. Our experiments across several GLUE and SuperGLUE benchmarks consistently highlight SUPERPOS-PROMPT's superiority over Residual Prompt tuning, exhibiting an average score increase of +6.4 in T5-Small and +5.0 in T5-Base along with a faster convergence. Remarkably, SUPERPOS-PROMPT occasionally outperforms even full fine-tuning methods. (ii) Additionally, we demonstrate enhanced performance and rapid convergence by omitting dropouts from the frozen network, yielding consistent improvements across various scenarios and tuning methods.

Original languageEnglish
Pages (from-to)34-46
Number of pages13
JournalProceedings of Machine Learning Research
Volume262
DOIs
Publication statusPublished - 7 Jun 2024
Event4th NeurIPS Efficient Natural Language and Speech Processing Workshop - Vancouver, Canada
Duration: 14 Dec 2024 → …

Fingerprint

Dive into the research topics of 'SuperPos-Prompt: Enhancing Soft Prompt Tuning of Language Models with Superposition of Multi Token Embeddings'. Together they form a unique fingerprint.

Cite this