SOTAVerified

Masked Language Modeling

Papers

Showing 301350 of 475 papers

TitleStatusHype
Misinformation Detection in Social Media Video Posts0
Self-Supervised Representation Learning for Speech Using Visual Grounding and Masked Language ModelingCode1
Prompt-Guided Injection of Conformation to Pre-trained Protein Model0
MVPTR: Multi-Level Semantic Alignment for Vision-Language Pre-Training via Multi-Stage LearningCode1
Text Style Transfer for Bias Mitigation using Masked Language Modeling0
Data Augmentation for Biomedical Factoid Question Answering0
Causal Distillation for Language Models0
STT: Soft Template Tuning for Few-Shot Learning0
Eliciting Knowledge from Pretrained Language Models for Prototypical Prompt VerbalizerCode1
A Transfer Learning Pipeline for Educational Resource Discovery with Application in Leading Paragraph Generation0
EvoMoE: An Evolutional Mixture-of-Experts Training Framework via Dense-To-Sparse GateCode1
Does Pre-training Induce Systematic Inference? How Masked Language Models Acquire Commonsense Knowledge0
CoCo-BERT: Improving Video-Language Pre-training with Contrastive Cross-modal Matching and Denoising0
Unified Multimodal Pre-training and Prompt-based Tuning for Vision-Language Understanding and Generation0
DIBERT: Dependency Injected Bidirectional Encoder Representations from TransformersCode0
Causal Distillation for Language ModelsCode1
UFO: A UniFied TransfOrmer for Vision-Language Representation Learning0
LAnoBERT: System Log Anomaly Detection based on BERT Masked Language Model0
Prompt-Learning for Fine-Grained Entity Typing0
TACO: Pre-training of Deep Transformers with Attention Convolution using Disentangled Positional Representation0
Phrase-aware Unsupervised Constituency Parsing0
Temporal Language Modeling for Short Text Document Classification with Transformers0
How does the pre-training objective affect what large language models learn about linguistic properties?0
DS-TOD: Efficient Domain Specialization for Task-Oriented Dialog0
Generative Prompt Tuning for Relation Classification0
Towards Unified Prompt Tuning for Few-shot Learning0
Contextual Representation Learning beyond Masked Language Modeling0
Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt Verbalizer for Text Classification0
"Is Whole Word Masking Always Better for Chinese BERT?": Probing on Chinese Grammatical Error Correction0
Predicting Attention Sparsity in Transformers0
Probing BERT’s priors with serial reproduction chains0
DAWSON: Data Augmentation using Weak Supervision On Natural Language0
Composable Sparse Fine-Tuning for Cross-Lingual Transfer0
A Good Prompt Is Worth Millions of Parameters: Low-resource Prompt-based Learning for Vision-Language Models0
Unsupervised Dependency Graph Network0
Joint Unsupervised and Supervised Training for Multilingual ASR0
iBOT: Image BERT Pre-Training with Online TokenizerCode1
NICT Kyoto Submission for the WMT’21 Quality Estimation Task: Multimetric Multilingual Pretraining for Critical Error Detection0
Modeling Mathematical Notation Semantics in Academic Papers0
JavaBERT: Training a transformer-based model for the Java programming languageCode0
NormFormer: Improved Transformer Pretraining with Extra Normalization0
A Good Prompt Is Worth Millions of Parameters: Low-resource Prompt-based Learning for Vision-Language ModelsCode1
DS-TOD: Efficient Domain Specialization for Task Oriented DialogCode0
Composable Sparse Fine-Tuning for Cross-Lingual TransferCode1
Maximizing Efficiency of Language Model Pre-training for Learning Representation0
Dict-BERT: Enhancing Language Model Pre-training with DictionaryCode0
Multi-Modal Pre-Training for Automated Speech Recognition0
Contextualized Semantic Distance between Highly Overlapped TextsCode0
Image BERT Pre-training with Online Tokenizer0
MLIM: Vision-and-Language Model Pre-training with Masked Language and Image Modeling0
Show:102550
← PrevPage 7 of 10Next →

No leaderboard results yet.