SOTAVerified

Masked Language Modeling

Papers

Showing 201250 of 475 papers

TitleStatusHype
Less is More: Pre-Training Cross-Lingual Small-Scale Language Models with Cognitively-Plausible Curriculum Learning StrategiesCode0
Leveraging Open Information Extraction for More Robust Domain Transfer of Event Trigger DetectionCode0
Lil-Bevo: Explorations of Strategies for Training Language Models in More Humanlike WaysCode0
Masked and Permuted Implicit Context Learning for Scene Text RecognitionCode0
Masked Language Modeling for Proteins via Linearly Scalable Long-Context TransformersCode0
Masked Language Models are Good Heterogeneous Graph GeneralizersCode0
Masked Latent Semantic Modeling: an Efficient Pre-training Alternative to Masked Language ModelingCode0
Mask-Enhanced Autoregressive Prediction: Pay Less Attention to Learn MoreCode0
Measuring Social Biases in Masked Language Models by Proxy of Prediction QualityCode0
MeLT: Message-Level Transformer with Masked Document Representations as Pre-Training for Stance DetectionCode0
Mistral-SPLADE: LLMs for better Learned Sparse RetrievalCode0
MSA TransformerCode0
Multilingual Normalization of Temporal Expressions with Masked Language ModelsCode0
Multilinguals at SemEval-2022 Task 11: Complex NER in Semantically Ambiguous Settings for Low Resource LanguagesCode0
NormFormer: Improved Transformer Pretraining with Extra NormalizationCode0
On the Cross-lingual Transferability of Monolingual RepresentationsCode0
PEACH: Pre-Training Sequence-to-Sequence Multilingual Models for Translation with Semi-Supervised Pseudo-Parallel Document GenerationCode0
Personalized Image Enhancement Featuring Masked Style ModelingCode0
Plausible May Not Be Faithful: Probing Object Hallucination in Vision-Language Pre-trainingCode0
Boosting Point-BERT by Multi-choice TokensCode0
Pre-Training of Deep Bidirectional Protein Sequence Representations with Structural InformationCode0
Pre-Training Transformer Decoder for End-to-End ASR Model with Unpaired Speech DataCode0
Pre-training with Aspect-Content Text Mutual Prediction for Multi-Aspect Dense RetrievalCode0
Probing BERT's priors with serial reproduction chainsCode0
Promises and Pitfalls of Generative Masked Language Modeling: Theoretical Framework and Practical GuidelinesCode0
PromptCL: Improving Event Representation via Prompt Template and Contrastive LearningCode0
Punctuation Restoration Improves Structure Understanding Without SupervisionCode0
QueerBench: Quantifying Discrimination in Language Models Toward Queer IdentitiesCode0
RAVEN: In-Context Learning with Retrieval-Augmented Encoder-Decoder Language ModelsCode0
ReCAM@IITK at SemEval-2021 Task 4: BERT and ALBERT based Ensemble for Abstract Word PredictionCode0
Revisiting and Advancing Chinese Natural Language Understanding with Accelerated Heterogeneous Knowledge Pre-trainingCode0
S2SNet: A Pretrained Neural Network for Superconductivity DiscoveryCode0
SAS: Self-Augmentation Strategy for Language Model Pre-trainingCode0
SciPrompt: Knowledge-augmented Prompting for Fine-grained Categorization of Scientific TopicsCode0
Self-Distillation Improves DNA Sequence InferenceCode0
Self-Evolution Learning for Discriminative Language Model PretrainingCode0
Selfie: Self-supervised Pretraining for Image EmbeddingCode0
Seventeenth-Century Spanish American Notary Records for Fine-Tuning Spanish Large Language ModelsCode0
SJ_AJ@DravidianLangTech-EACL2021: Task-Adaptive Pre-Training of Multilingual BERT models for Offensive Language IdentificationCode0
StrucTexTv2: Masked Visual-Textual Prediction for Document Image Pre-trainingCode0
Structural Self-Supervised Objectives for TransformersCode0
Substrate Prediction for RiPP Biosynthetic Enzymes via Masked Language Modeling and Transfer LearningCode0
Symbolic Discovery of Optimization AlgorithmsCode0
Task-Informed Anti-Curriculum by Masking Improves Downstream Performance on TextCode0
Text Revision by On-the-Fly Representation OptimizationCode0
The Distributional Hypothesis Does Not Fully Explain the Benefits of Masked Language Model PretrainingCode0
The Effectiveness of Masked Language Modeling and Adapters for Factual Knowledge InjectionCode0
Towards Unified Prompt Tuning for Few-shot Text ClassificationCode0
Towards Unifying Reference Expression Generation and ComprehensionCode0
Transformer based neural networks for emotion recognition in conversationsCode0
Show:102550
← PrevPage 5 of 10Next →

No leaderboard results yet.