TY - JOUR
T1 - SuperPos-Prompt
T2 - 4th NeurIPS Efficient Natural Language and Speech Processing Workshop
AU - SadraeiJavaeri, Mohammad Ali
AU - Asgari, Ehsaneddin
AU - McHardy, Alice Carolyn
AU - Rabiee, Hamid Reza
N1 - Publisher Copyright:
© 2024 Proceedings of Machine Learning Research.
PY - 2024/6/7
Y1 - 2024/6/7
N2 - Soft prompt tuning techniques have recently gained traction as an effective strategy for the parameter-efficient tuning of pretrained language models, particularly minimizing the required adjustment of model parameters. Despite their growing use, achieving optimal tuning with soft prompts, especially for smaller datasets, remains a substantial challenge. This study makes two contributions in this domain: (i) we introduce SUPERPOS-PROMPT, a new reparameterization technique employing the superposition of multiple pretrained vocabulary embeddings to improve the learning of soft prompts. Our experiments across several GLUE and SuperGLUE benchmarks consistently highlight SUPERPOS-PROMPT's superiority over Residual Prompt tuning, exhibiting an average score increase of +6.4 in T5-Small and +5.0 in T5-Base along with a faster convergence. Remarkably, SUPERPOS-PROMPT occasionally outperforms even full fine-tuning methods. (ii) Additionally, we demonstrate enhanced performance and rapid convergence by omitting dropouts from the frozen network, yielding consistent improvements across various scenarios and tuning methods.
AB - Soft prompt tuning techniques have recently gained traction as an effective strategy for the parameter-efficient tuning of pretrained language models, particularly minimizing the required adjustment of model parameters. Despite their growing use, achieving optimal tuning with soft prompts, especially for smaller datasets, remains a substantial challenge. This study makes two contributions in this domain: (i) we introduce SUPERPOS-PROMPT, a new reparameterization technique employing the superposition of multiple pretrained vocabulary embeddings to improve the learning of soft prompts. Our experiments across several GLUE and SuperGLUE benchmarks consistently highlight SUPERPOS-PROMPT's superiority over Residual Prompt tuning, exhibiting an average score increase of +6.4 in T5-Small and +5.0 in T5-Base along with a faster convergence. Remarkably, SUPERPOS-PROMPT occasionally outperforms even full fine-tuning methods. (ii) Additionally, we demonstrate enhanced performance and rapid convergence by omitting dropouts from the frozen network, yielding consistent improvements across various scenarios and tuning methods.
UR - https://www.scopus.com/pages/publications/85216733789
U2 - 10.48550/arXiv.2406.05279
DO - 10.48550/arXiv.2406.05279
M3 - Conference article
AN - SCOPUS:85216733789
SN - 2640-3498
VL - 262
SP - 34
EP - 46
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
Y2 - 14 December 2024
ER -