SOTAVerified

Masked Language Modeling

Papers

Showing 151200 of 475 papers

TitleStatusHype
LayoutMask: Enhance Text-Layout Interaction in Multi-modal Pre-training for Document Understanding0
Preserving Pre-trained Features Helps Calibrate Fine-tuned Language ModelsCode1
Adapting Learned Sparse Retrieval for Long DocumentsCode0
Rethinking Masked Language Modeling for Chinese Spelling CorrectionCode1
Honey, I Shrunk the Language: Language Model Behavior at Reduced ScaleCode0
An Investigation of Noise in Morphological InflectionCode0
Masked and Permuted Implicit Context Learning for Scene Text RecognitionCode0
Dynamic Masking Rate Schedules for MLM Pretraining0
Self-Evolution Learning for Discriminative Language Model PretrainingCode0
Leveraging Open Information Extraction for More Robust Domain Transfer of Event Trigger DetectionCode0
AxomiyaBERTa: A Phonologically-aware Transformer Model for AssameseCode0
Bidirectional Transformer Reranker for Grammatical Error CorrectionCode0
Extrapolating Multilingual Understanding Models as Multilingual Generators0
Federated Learning of Medical Concepts Embedding using BEHRTCode0
A Pilot Study on Dialogue-Level Dependency Parsing for Chinese0
ESCOXLM-R: Multilingual Taxonomy-driven Pre-training for the Job Market DomainCode1
Patton: Language Model Pretraining on Text-Rich Networks0
Syllable Discovery and Cross-Lingual Generalization in a Visually Grounded, Self-Supervised Speech ModelCode1
How does the task complexity of masked pretraining objectives affect downstream performance?Code0
DinoSR: Self-Distillation and Online Clustering for Self-supervised Speech Representation LearningCode1
Pre-training Language Model as a Multi-perspective Course Learner0
Mapping of attention mechanisms to a generalized Potts model0
Unsupervised Improvement of Factual Knowledge in Language ModelsCode0
PEACH: Pre-Training Sequence-to-Sequence Multilingual Models for Translation with Semi-Supervised Pseudo-Parallel Document GenerationCode0
Joint unsupervised and supervised learning for context-aware language identification0
Fine-grained Audible Video DescriptionCode1
Accelerating Vision-Language Pretraining with Free Language ModelingCode1
Cross-Modal Implicit Relation Reasoning and Aligning for Text-to-Image Person RetrievalCode2
HOP+: History-enhanced and Order-aware Pre-training for Vision-and-Language Navigation0
CCPL: Cross-modal Contrastive Protein Learning0
Do Transformers Parse while Predicting the Masked Word?0
Generating multiple-choice questions for medical question answering with distractors and cue-masking0
Domain-adapted large language models for classifying nuclear medicine reports0
StrucTexTv2: Masked Visual-Textual Prediction for Document Image Pre-trainingCode0
Efficient Masked Autoencoders with Self-Consistency0
Weighted Sampling for Masked Language Modeling0
Symbolic Discovery of Optimization AlgorithmsCode0
Capturing Topic Framing via Masked Language Modeling0
Representation Deficiency in Masked Language ModelingCode1
Tagging before Alignment: Integrating Multi-Modal Tags for Video-Text Retrieval0
Debiasing the Cloze Task in Sequential Recommendation with Bidirectional TransformersCode1
A Cohesive Distillation Architecture for Neural Language Models0
Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks0
Cramming: Training a Language Model on a Single GPU in One DayCode3
MicroBERT: Effective Training of Low-resource Monolingual BERTs through Parameter Reduction and Multitask LearningCode1
Go-tuning: Improving Zero-shot Learning Abilities of Smaller Language Models0
Mu^2SLAM: Multitask, Multilingual Speech and Language Models0
APOLLO: A Simple Approach for Adaptive Pretraining of Language Models for Logical Reasoning0
Efficient Pre-training of Masked Language Model via Concept-based Curriculum MaskingCode1
Uniform Masking Prevails in Vision-Language Pretraining0
Show:102550
← PrevPage 4 of 10Next →

No leaderboard results yet.