SOTAVerified

Protum: A New Method For Prompt Tuning Based on "[MASK]"

2022-01-28Unverified0· sign in to hype

Pan He, Yuxi Chen, Yan Wang, Yanru Zhang

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Recently, prompt tuning lester2021power has gradually become a new paradigm for NLP, which only depends on the representation of the words by freezing the parameters of pre-trained language models (PLMs) to obtain remarkable performance on downstream tasks. It maintains the consistency of Masked Language Model (MLM) devlin2018bert task in the process of pre-training, and avoids some issues that may happened during fine-tuning. Naturally, we consider that the "[MASK]" tokens carry more useful information than other tokens because the model combines with context to predict the masked tokens. Among the current prompt tuning methods, there will be a serious problem of random composition of the answer tokens in prediction when they predict multiple words so that they have to map tokens to labels with the help verbalizer. In response to the above issue, we propose a new Prompt Tuning based on "[MASK]" (Protum) method in this paper, which constructs a classification task through the information carried by the hidden layer of "[MASK]" tokens and then predicts the labels directly rather than the answer tokens. At the same time, we explore how different hidden layers under "[MASK]" impact on our classification model on many different data sets. Finally, we find that our Protum can achieve much better performance than fine-tuning after continuous pre-training with less time consumption. Our model facilitates the practical application of large models in NLP.

Tasks

Reproductions