SOTAVerified

Masked Language Modeling

Papers

Showing 426450 of 475 papers

TitleStatusHype
SPLAT: Speech-Language Joint Pre-Training for Spoken Language UnderstandingCode1
XDA: Accurate, Robust Disassembly with Transfer LearningCode1
GraPPa: Grammar-Augmented Pre-Training for Table Semantic ParsingCode1
VECO: Variable Encoder-decoder Pre-training for Cross-lingual Understanding and Generation0
Deep Transformers with Latent DepthCode0
GraphCodeBERT: Pre-training Code Representations with Data Flow0
Intermediate Training of BERT for Product MatchingCode1
Learning Visual Representations with Caption Annotations0
TensorCoder: Dimension-Wise Attention via Tensor Representation for Natural Language Modeling0
The Lottery Ticket Hypothesis for Pre-trained BERT NetworksCode1
Language-agnostic BERT Sentence EmbeddingCode1
Pre-training via ParaphrasingCode1
I-BERT: Inductive Generalization of Transformer to Arbitrary Context LengthsCode0
MC-BERT: Efficient Language Pre-Training via a Meta ControllerCode1
GMAT: Global Memory Augmentation for TransformersCode0
Masked Language Modeling for Proteins via Linearly Scalable Long-Context Transformers0
Segatron: Segment-aware Transformer for Language Modeling and Understanding0
Position Masking for Language Models0
Massive Choice, Ample Tasks (MaChAmp): A Toolkit for Multi-task Learning in NLPCode1
HERO: Hierarchical Encoder for Video+Language Omni-representation Pre-trainingCode1
Segatron: Segment-Aware Transformer for Language Modeling and UnderstandingCode1
Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning0
UHH-LT at SemEval-2020 Task 12: Fine-Tuning of Pre-Trained Transformer Networks for Offensive Language Detection0
Train No Evil: Selective Masking for Task-Guided Pre-TrainingCode1
MPNet: Masked and Permuted Pre-training for Language UnderstandingCode2
Show:102550
← PrevPage 18 of 19Next →

No leaderboard results yet.