SuperPos-Prompt: Enhancing Soft Prompt Tuning of Language Models with Superposition of Multi Token EmbeddingsDownload PDF

Anonymous

16 Feb 2024 (modified: 16 Dec 2024)ACL ARR 2024 February Blind SubmissionReaders: Everyone
Abstract: Soft prompt tuning techniques have recently gained traction as an effective strategy for the parameter-efficient tuning of pretrained language models, particularly minimizing the required adjustment of model parameters. Despite their growing use, achieving optimal tuning with soft prompts, especially with smaller datasets, remains a substantial challenge. This study makes two contributions in this domain: (i) we introduce \textsc{SuperPos-Prompt}, a new reparameterization technique employing the superposition of multiple pretrained vocabulary embeddings to improve the learning of soft prompts. Our experiments across several GLUE and SuperGLUE benchmarks consistently highlight \textsc{SuperPos-Prompt}'s superiority over \textit{Residual Prompt} tuning, exhibiting an average score increase of $+6.4$ in \textit{T5-Small} and $+5.0$ in \textit{T5-Base} along with a faster convergence. Remarkably, \textsc{SuperPos-Prompt} occasionally outperforms even full fine-tuning methods. (ii) Additionally, we demonstrate enhanced performance and rapid convergence by omitting dropout from the frozen network, yielding consistent improvements across various scenarios and tuning methods.
Paper Type: long
Research Area: Efficient/Low-Resource Methods for NLP
Contribution Types: Model analysis & interpretability, NLP engineering experiment, Reproduction study, Approaches low compute settings-efficiency
Languages Studied: English
0 Replies

Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview