SOTAVerified

Masked Language Modeling

Papers

Showing 401450 of 475 papers

TitleStatusHype
MSA Transformer0
SJ_AJ@DravidianLangTech-EACL2021: Task-Adaptive Pre-Training of Multilingual BERT models for Offensive Language IdentificationCode0
MERMAID: Metaphor Generation with Symbolism and Discriminative Decoding0
CDLM: Cross-Document Language ModelingCode1
Universal Sentence Representations Learning with Conditional Masked Language Model0
AraELECTRA: Pre-Training Text Discriminators for Arabic Language UnderstandingCode1
Universal Sentence Representation Learning with Conditional Masked Language Model0
RealFormer: Transformer Likes Residual AttentionCode1
TAP: Text-Aware Pre-training for Text-VQA and Text-CaptionCode1
Pre-training Protein Language Models with Label-Agnostic Binding Pairs Enhances Performance in Downstream TasksCode1
XHate-999: Analyzing and Detecting Abusive Language Across Domains and Languages0
StructFormer: Joint Unsupervised Induction of Dependency and Constituency Structure from Masked Language ModelingCode1
Profile Prediction: An Alignment-Based Pre-Training Task for Protein Sequence Models0
Self-Supervised Relationship Probing0
Self-Supervised learning with cross-modal transformers for emotion recognition0
A Hierarchical Multi-Modal Encoder for Moment Localization in Video Corpus0
POSTECH-ETRI’s Submission to the WMT2020 APE Shared Task: Automatic Post-Editing with Cross-lingual Language Model0
Controlling the Imprint of Passivization and Negation in Contextualized RepresentationsCode0
Effective Decoder Masking for Transformer Based End-to-End Speech Recognition0
DICT-MLM: Improved Multilingual Pre-Training using Bilingual Dictionaries0
ST-BERT: Cross-modal Language Model Pre-training For End-to-end Spoken Language Understanding0
ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling for Natural Language UnderstandingCode3
Cold-start Active Learning through Self-supervised Language ModelingCode1
Corruption Is Not All Bad: Incorporating Discourse Structure into Pre-training via Corruption for Essay Scoring0
Cross-Thought for Sentence Encoder Pre-trainingCode1
SPLAT: Speech-Language Joint Pre-Training for Spoken Language UnderstandingCode1
XDA: Accurate, Robust Disassembly with Transfer LearningCode1
GraPPa: Grammar-Augmented Pre-Training for Table Semantic ParsingCode1
VECO: Variable Encoder-decoder Pre-training for Cross-lingual Understanding and Generation0
Deep Transformers with Latent DepthCode0
GraphCodeBERT: Pre-training Code Representations with Data Flow0
Intermediate Training of BERT for Product MatchingCode1
Learning Visual Representations with Caption Annotations0
TensorCoder: Dimension-Wise Attention via Tensor Representation for Natural Language Modeling0
The Lottery Ticket Hypothesis for Pre-trained BERT NetworksCode1
Language-agnostic BERT Sentence EmbeddingCode1
Pre-training via ParaphrasingCode1
I-BERT: Inductive Generalization of Transformer to Arbitrary Context LengthsCode0
MC-BERT: Efficient Language Pre-Training via a Meta ControllerCode1
GMAT: Global Memory Augmentation for TransformersCode0
Masked Language Modeling for Proteins via Linearly Scalable Long-Context Transformers0
Segatron: Segment-aware Transformer for Language Modeling and Understanding0
Position Masking for Language Models0
Massive Choice, Ample Tasks (MaChAmp): A Toolkit for Multi-task Learning in NLPCode1
HERO: Hierarchical Encoder for Video+Language Omni-representation Pre-trainingCode1
Segatron: Segment-Aware Transformer for Language Modeling and UnderstandingCode1
Pre-training Is (Almost) All You Need: An Application to Commonsense Reasoning0
UHH-LT at SemEval-2020 Task 12: Fine-Tuning of Pre-Trained Transformer Networks for Offensive Language Detection0
Train No Evil: Selective Masking for Task-Guided Pre-TrainingCode1
MPNet: Masked and Permuted Pre-training for Language UnderstandingCode2
Show:102550
← PrevPage 9 of 10Next →

No leaderboard results yet.