SOTAVerified

CMV-BERT: Contrastive multi-vocab pretraining of BERT

2020-12-29Unverified0· sign in to hype

Wei Zhu, Daniel Cheung

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In this work, we represent CMV-BERT, which improves the pretraining of a language model via two ingredients: (a) contrastive learning, which is well studied in the area of computer vision; (b) multiple vocabularies, one of which is fine-grained and the other is coarse-grained. The two methods both provide different views of an original sentence, and both are shown to be beneficial. Downstream tasks demonstrate our proposed CMV-BERT are effective in improving the pretrained language models.

Tasks

Reproductions