SOTAVerified

Masked Language Modeling

Papers

Showing 351400 of 475 papers

TitleStatusHype
Predicting Attention Sparsity in Transformers0
MeLT: Message-Level Transformer with Masked Document Representations as Pre-Training for Stance DetectionCode0
SupCL-Seq: Supervised Contrastive Learning for Downstream Optimized Sequence RepresentationsCode1
CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and GenerationCode1
Data Efficient Masked Language Modeling for Vision and LanguageCode1
Frustratingly Simple Pretraining Alternatives to Masked Language ModelingCode1
Split-and-Rephrase in a Cross-Lingual Manner: A Complete Pipeline0
Domain-Specific Japanese ELECTRA Model Using a Small Corpus0
CTAL: Pre-training Cross-modal Transformer for Audio-and-Language RepresentationsCode1
Sentence Bottleneck Autoencoders from Transformer Language ModelsCode1
MELM: Data Augmentation with Masked Entity Language Modeling for Low-Resource NERCode1
Prompt-Learning for Fine-Grained Entity Typing0
Knowledge Perceived Multi-modal Pretraining in E-commerceCode1
W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-TrainingCode3
Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt Verbalizer for Text ClassificationCode1
Noobs at Semeval-2021 Task 4: Masked Language Modeling for abstract answer prediction0
Fine-Grained Emotion Prediction by Modeling Emotion DefinitionsCode0
Learning to Sample Replacements for ELECTRA Pre-Training0
Winner Team Mia at TextVQA Challenge 2021: Vision-and-Language Representation Learning with Pre-trained Sequence-to-Sequence Model0
SPBERT: An Efficient Pre-training BERT on SPARQL Queries for Question Answering over Knowledge GraphsCode1
SAS: Self-Augmentation Strategy for Language Model Pre-trainingCode0
Improving Pretrained Cross-Lingual Language Models via Self-Labeled Word AlignmentCode1
Exploring Unsupervised Pretraining Objectives for Machine TranslationCode0
MST: Masked Self-Supervised Transformer for Visual Representation0
BERTnesia: Investigating the capture and forgetting of knowledge in BERTCode0
Exposing the Implicit Energy Networks behind Masked Language Models via Metropolis--Hastings0
Bi-Granularity Contrastive Learning for Post-Training in Few-Shot Scene0
Luna: Linear Unified Nested AttentionCode1
BERT-Defense: A Probabilistic Model Based on BERT to Combat Cognitively Inspired Orthographic Adversarial AttacksCode0
MG-BERT: Multi-Graph Augmented BERT for Masked Language Modeling0
SCRIPT: Self-Critic PreTraining of Transformers0
Target-Aware Data Augmentation for Stance Detection0
TreeBERT: A Tree-Based Pre-Trained Model for Programming LanguageCode1
From Masked Language Modeling to Translation: Non-English Auxiliary Tasks Improve Zero-shot Spoken Language UnderstandingCode0
Larger-Scale Transformers for Multilingual Masked Language Modeling0
Understanding Chinese Video and Language via Contrastive Multimodal Pre-Training0
On the Influence of Masking Policies in Intermediate Pre-training0
KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation ExtractionCode1
Masked Language Modeling and the Distributional Hypothesis: Order Word Matters Pre-training for Little0
On the Inductive Bias of Masked Language Modeling: From Statistical to Syntactic DependenciesCode1
ReCAM@IITK at SemEval-2021 Task 4: BERT and ALBERT based Ensemble for Abstract Word PredictionCode0
MMBERT: Multimodal BERT Pretraining for Improved Medical VQACode1
Pseudo-Label Guided Unsupervised Domain Adaptation of Contextual Embeddings0
UC2: Universal Cross-lingual Cross-modal Vision-and-Language Pre-training0
Self-supervised Image-text Pre-training With Mixed Data In Chest X-rays0
Variable Name Recovery in Decompiled Binary Code using Constrained Masked Language Modeling0
Improving the Lexical Ability of Pretrained Language Models for Unsupervised Neural Machine TranslationCode1
MERMAID: Metaphor Generation with Symbolism and Discriminative DecodingCode1
A Primer on Contrastive Pretraining in Language Processing: Methods, Lessons Learned and Perspectives0
Bilingual Language Modeling, A transfer learning technique for Roman Urdu0
Show:102550
← PrevPage 8 of 10Next →

No leaderboard results yet.