SOTAVerified

Masked Language Modeling

Papers

Showing 201250 of 475 papers

TitleStatusHype
Embracing Ambiguity: Improving Similarity-oriented Tasks with Contextual Synonym Knowledge0
Emerging Cross-lingual Structure in Pretrained Language Models0
Emerging Property of Masked Token for Effective Pre-training0
Enabling Autoregressive Models to Fill In Masked Tokens0
Enhancing BERT-Based Visual Question Answering through Keyword-Driven Sentence Selection0
Enhancing Domain-Specific Encoder Models with LLM-Generated Data: How to Leverage Ontologies, and How to Do Without Them0
ESALE: Enhancing Code-Summary Alignment Learning for Source Code Summarization0
Exploring Multi-Modal Contextual Knowledge for Open-Vocabulary Object Detection0
Exposing the Implicit Energy Networks behind Masked Language Models via Metropolis--Hastings0
Extrapolating Multilingual Understanding Models as Multilingual Generators0
FARM: Functional Group-Aware Representations for Small Molecules0
How Useful is Continued Pre-Training for Generative Unsupervised Domain Adaptation?0
Foundation Posteriors for Approximate Probabilistic Inference0
Gender-tuning: Empowering Fine-tuning for Debiasing Pre-trained Language Models0
General Framework for Reversible Data Hiding in Texts Based on Masked Language Modeling0
Generating multiple-choice questions for medical question answering with distractors and cue-masking0
Generative Prompt Tuning for Relation Classification0
GeoRecon: Graph-Level Representation Learning for 3D Molecules via Reconstruction-Based Pretraining0
Global memory transformer for processing long documents0
Go-tuning: Improving Zero-shot Learning Abilities of Smaller Language Models0
GPTs at Factify 2022: Prompt Aided Fact-Verification0
GraphCodeBERT: Pre-training Code Representations with Data Flow0
HAD: Hybrid Architecture Distillation Outperforms Teacher in Genomic Sequence Modeling0
HCDIR: End-to-end Hate Context Detection, and Intensity Reduction model for online comments0
HOP+: History-enhanced and Order-aware Pre-training for Vision-and-Language Navigation0
How does the pre-training objective affect what large language models learn about linguistic properties?0
Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks0
ImageBERT: Cross-modal Pre-training with Large-scale Weak-supervised Image-Text Data0
Image BERT Pre-training with Online Tokenizer0
Improving BERT with Hybrid Pooling Network and Drop Mask0
Improving Low-Resource Morphological Inflection via Self-Supervised Objectives0
Improving the Reusability of Pre-trained Language Models in Real-world Applications0
In-Context Learning can distort the relationship between sequence likelihoods and biological fitness0
Investigating Masking-based Data Generation in Language Models0
"Is Whole Word Masking Always Better for Chinese BERT?": Probing on Chinese Grammatical Error Correction0
"Is Whole Word Masking Always Better for Chinese BERT?": Probing on Chinese Grammatical Error Correction0
“Is Whole Word Masking Always Better for Chinese BERT?”: Probing on Chinese Grammatical Error Correction0
Iterative Mask Filling: An Effective Text Augmentation Method Using Masked Language Modeling0
Joint unsupervised and supervised learning for context-aware language identification0
Joint Unsupervised and Supervised Training for Multilingual ASR0
KECP: Knowledge Enhanced Contrastive Prompting for Few-shot Extractive Question Answering0
Knowing Where to Focus: Attention-Guided Alignment for Text-based Person Search0
Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt Verbalizer for Text Classification0
Knowledge Distillation vs. Pretraining from Scratch under a Fixed (Computation) Budget0
KUL@SMM4H’22: Template Augmented Adaptive Pre-training for Tweet Classification0
LakotaBERT: A Transformer-based Model for Low Resource Lakota Language0
LAnoBERT: System Log Anomaly Detection based on BERT Masked Language Model0
Larger-Scale Transformers for Multilingual Masked Language Modeling0
LayoutMask: Enhance Text-Layout Interaction in Multi-modal Pre-training for Document Understanding0
Enhancing Continual Learning with Global Prototypes: Counteracting Negative Representation Drift0
Show:102550
← PrevPage 5 of 10Next →

No leaderboard results yet.