SOTAVerified

Masked Language Modeling

Papers

Showing 401450 of 475 papers

TitleStatusHype
JavaBERT: Training a transformer-based model for the Java programming languageCode0
Dict-BERT: Enhancing Language Model Pre-training with DictionaryCode0
Boosting Prompt-Based Self-Training With Mapping-Free Automatic Verbalizer for Multi-Class ClassificationCode0
Specializing Unsupervised Pretraining Models for Word-Level Semantic SimilarityCode0
Biomedical Language Models are Robust to Sub-optimal TokenizationCode0
Bidirectional Transformer Reranker for Grammatical Error CorrectionCode0
DIBERT: Dependency Injected Bidirectional Encoder Representations from TransformersCode0
On the Cross-lingual Transferability of Monolingual RepresentationsCode0
Image as a Foreign Language: BEiT Pretraining for All Vision and Vision-Language TasksCode0
IDIAPers @ Causal News Corpus 2022: Efficient Causal Relation Identification Through a Prompt-based Few-shot ApproachCode0
Transformer based neural networks for emotion recognition in conversationsCode0
BERTnesia: Investigating the capture and forgetting of knowledge in BERTCode0
StrucTexTv2: Masked Visual-Textual Prediction for Document Image Pre-trainingCode0
I-BERT: Inductive Generalization of Transformer to Arbitrary Context LengthsCode0
Deep Transformers with Latent DepthCode0
Structural Self-Supervised Objectives for TransformersCode0
PEACH: Pre-Training Sequence-to-Sequence Multilingual Models for Translation with Semi-Supervised Pseudo-Parallel Document GenerationCode0
How transformers learn structured data: insights from hierarchical filteringCode0
BERT-Defense: A Probabilistic Model Based on BERT to Combat Cognitively Inspired Orthographic Adversarial AttacksCode0
Personalized Image Enhancement Featuring Masked Style ModelingCode0
AxomiyaBERTa: A Phonologically-aware Transformer Model for AssameseCode0
Substrate Prediction for RiPP Biosynthetic Enzymes via Masked Language Modeling and Transfer LearningCode0
Plausible May Not Be Faithful: Probing Object Hallucination in Vision-Language Pre-trainingCode0
Boosting Point-BERT by Multi-choice TokensCode0
How does the task complexity of masked pretraining objectives affect downstream performance?Code0
Data Augmentation for Biomedical Factoid Question AnsweringCode0
Counterfactually Probing Language Identity in Multilingual ModelsCode0
Symbolic Discovery of Optimization AlgorithmsCode0
Contextualized Semantic Distance between Highly Overlapped TextsCode0
Honey, I Shrunk the Language: Language Model Behavior at Reduced ScaleCode0
Historical Ink: Semantic Shift Detection for 19th Century SpanishCode0
Unsupervised Domain Adaptation of Contextualized Embeddings for Sequence LabelingCode0
Unsupervised Improvement of Factual Knowledge in Language ModelsCode0
An Investigation of Noise in Morphological InflectionCode0
Controlling the Imprint of Passivization and Negation in Contextualized RepresentationsCode0
Pre-Training of Deep Bidirectional Protein Sequence Representations with Structural InformationCode0
HanTrans: An Empirical Study on Cross-Era Transferability of Chinese Pre-trained Language ModelCode0
Dial-MAE: ConTextual Masked Auto-Encoder for Retrieval-based Dialogue SystemsCode0
GMAT: Global Memory Augmentation for TransformersCode0
Pre-training with Aspect-Content Text Mutual Prediction for Multi-Aspect Dense RetrievalCode0
Probing BERT's priors with serial reproduction chainsCode0
Unsupervised Representation Learning of Player Behavioral Data with Confidence Guided MaskingCode0
Task-Informed Anti-Curriculum by Masking Improves Downstream Performance on TextCode0
MMCLIP: Cross-modal Attention Masked Modelling for Medical Language-Image Pre-TrainingCode0
PromptCL: Improving Event Representation via Prompt Template and Contrastive LearningCode0
A character-based steganography using masked language modelingCode0
ASMA-Tune: Unlocking LLMs' Assembly Code Comprehension via Structural-Semantic Instruction TuningCode0
Text Revision by On-the-Fly Representation OptimizationCode0
An Empirical Study Of Self-supervised Learning Approaches For Object Detection With TransformersCode0
AllenNLP Interpret: A Framework for Explaining Predictions of NLP ModelsCode0
Show:102550
← PrevPage 9 of 10Next →

No leaderboard results yet.