SOTAVerified

Masked Language Modeling

Papers

Showing 150 of 475 papers

TitleStatusHype
GigaAM: Efficient Self-Supervised Learner for Speech RecognitionCode4
Simple and Effective Masked Diffusion Language ModelsCode4
Towards No.1 in CLUE Semantic Matching Challenge: Pre-trained Language Model Erlangshen with Propensity-Corrected LossCode4
GLIPv2: Unifying Localization and Vision-Language UnderstandingCode4
Long-VITA: Scaling Large Multi-modal Models to 1 Million Tokens with Leading Short-Context AccurayCode3
Cramming: Training a Language Model on a Single GPU in One DayCode3
W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-TrainingCode3
ERNIE-Gram: Pre-Training with Explicitly N-Gram Masked Language Modeling for Natural Language UnderstandingCode3
BMFM-RNA: An Open Framework for Building and Evaluating Transcriptomic Foundation ModelsCode2
GPT or BERT: why not both?Code2
MosaicBERT: A Bidirectional Encoder Optimized for Fast PretrainingCode2
Cross-Modal Implicit Relation Reasoning and Aligning for Text-to-Image Person RetrievalCode2
Retrieval Oriented Masking Pre-training Language Model for Dense Passage RetrievalCode2
Deep Bidirectional Language-Knowledge Graph PretrainingCode2
RetroMAE: Pre-Training Retrieval-oriented Language Models Via Masked Auto-EncoderCode2
LinkBERT: Pretraining Language Models with Document LinksCode2
MPNet: Masked and Permuted Pre-training for Language UnderstandingCode2
Self-Supervised Log ParsingCode2
Diffusion Sequence Models for Enhanced Protein Representation and GenerationCode1
CreoPep: A Universal Deep Learning Framework for Target-Specific Peptide Design and OptimizationCode1
Long-context Protein Language Modeling Using Bidirectional Mamba with Shared Projection LayersCode1
DomURLs_BERT: Pre-trained BERT-based Model for Malicious Domains and URLs Detection and ClassificationCode1
AutoScale: Scale-Aware Data Mixing for Pre-Training LLMsCode1
Retrieval-style In-Context Learning for Few-shot Hierarchical Text ClassificationCode1
VCR: A Task for Pixel-Level Complex Reasoning in Vision Language Models via Restoring Occluded TextCode1
Merging Text Transformer Models from Different InitializationsCode1
NextLevelBERT: Masked Language Modeling with Higher-Level Representations for Long DocumentsCode1
CodeArt: Better Code Models by Attention Regularization When Symbols Are LackingCode1
Interpretation of Intracardiac Electrograms Through Textual RepresentationsCode1
Endowing Protein Language Models with Structural KnowledgeCode1
ECAMP: Entity-centered Context-aware Medical Vision Language Pre-trainingCode1
Labrador: Exploring the Limits of Masked Language Modeling for Laboratory DataCode1
GeoLM: Empowering Language Models for Geospatially Grounded Language UnderstandingCode1
FATA-Trans: Field And Time-Aware Transformer for Sequential Tabular DataCode1
FiLM: Fill-in Language Models for Any-Order GenerationCode1
PepMLM: Target Sequence-Conditioned Generation of Therapeutic Peptide Binders via Span Masked Language ModelingCode1
A Multi-Task Semantic Decomposition Framework with Task-specific Pre-training for Few-Shot NERCode1
Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-FinetuningCode1
Pairing interacting protein sequences using masked language modelingCode1
Stochastic positional embeddings improve masked image modelingCode1
Pushing the Limits of Unsupervised Unit Discovery for SSL Speech RepresentationCode1
Generate to Understand for RepresentationCode1
Global and Local Semantic Completion Learning for Vision-Language Pre-trainingCode1
On the Difference of BERT-style and CLIP-style Text EncodersCode1
Preserving Pre-trained Features Helps Calibrate Fine-tuned Language ModelsCode1
Rethinking Masked Language Modeling for Chinese Spelling CorrectionCode1
ESCOXLM-R: Multilingual Taxonomy-driven Pre-training for the Job Market DomainCode1
Syllable Discovery and Cross-Lingual Generalization in a Visually Grounded, Self-Supervised Speech ModelCode1
DinoSR: Self-Distillation and Online Clustering for Self-supervised Speech Representation LearningCode1
Fine-grained Audible Video DescriptionCode1
Show:102550
← PrevPage 1 of 10Next →

No leaderboard results yet.