SOTAVerified

Language Modeling

Papers

Showing 26012650 of 14182 papers

TitleStatusHype
Efficient Hierarchical Domain Adaptation for Pretrained Language ModelsCode1
Value Retrieval with Arbitrary Queries for Form-like DocumentsCode1
Improving Conversational Recommendation Systems' Quality with Context-Aware Item Meta InformationCode1
Deciphering antibody affinity maturation with language models and weakly supervised learningCode1
Step-unrolled Denoising Autoencoders for Text GenerationCode1
MAGMA -- Multimodal Augmentation of Generative Models through Adapter-based FinetuningCode1
MLP Architectures for Vision-and-Language Modeling: An Empirical StudyCode1
Zero-Shot Recommendation as Language ModelingCode1
Quantifying Adaptability in Pre-trained Language Models with 500 TasksCode1
Causal Distillation for Language ModelsCode1
Siamese BERT-based Model for Web Search Relevance Ranking Evaluated on a New Czech DatasetCode1
InfoLM: A New Metric to Evaluate Summarization & Data2Text GenerationCode1
Pixelated Butterfly: Simple and Efficient Sparse training for Neural Network ModelsCode1
ZeroCap: Zero-Shot Image-to-Text Generation for Visual-Semantic ArithmeticCode1
A Simple Long-Tailed Recognition Baseline via Vision-Language ModelCode1
Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered by Pre-Trained Vision-Language ModelCode1
UniTAB: Unifying Text and Box Outputs for Grounded Vision-Language ModelingCode1
Enhancing Multilingual Language Model with Massive Multilingual Knowledge TriplesCode1
RoBERTuito: a pre-trained language model for social media text in SpanishCode1
iBOT: Image BERT Pre-Training with Online TokenizerCode1
NLP From Scratch Without Large-Scale Pretraining: A Simple and Efficient FrameworkCode1
Tip-Adapter: Training-free CLIP-Adapter for Better Vision-Language ModelingCode1
A Pilot Study for BERT Language Modelling and Morphological Analysis for Ancient and Medieval GreekCode1
AESOP: Paraphrase Generation with Adaptive Syntactic ControlCode1
With a Little Help from my Temporal Context: Multimodal Egocentric Action RecognitionCode1
A Model of Cross-Lingual Knowledge-Grounded Response Generation for Open-Domain Dialogue SystemsCode1
Less is More: Pretrain a Strong Siamese Encoder for Dense Text Retrieval Using a Weak DecoderCode1
Small Data? No Problem! Exploring the Viability of Pretrained Multilingual Language Models for Low-resourced LanguagesCode1
AttentionRank: Unsupervised Keyphrase Extraction using Self and Cross AttentionsCode1
Effective Use of Graph Convolution Network and Contextual Sub-Tree for Commodity News Event ExtractionCode1
TSDAE: Using Transformer-based Sequential Denoising Auto-Encoderfor Unsupervised Sentence Embedding LearningCode1
Efficiently Modeling Long Sequences with Structured State SpacesCode1
Top1 Solution of QQ Browser 2021 Ai Algorithm Competition Track 1 : Multimodal Video SimilarityCode1
Scatterbrain: Unifying Sparse and Low-rank Attention ApproximationCode1
ÚFAL at MultiLexNorm 2021: Improving Multilingual Lexical Normalization by Fine-tuning ByT5Code1
Discovering Non-monotonic Autoregressive Orderings with Variational InferenceCode1
Deciphering the Language of Nature: A transformer-based language model for deleterious mutations in proteinsCode1
Hierarchical Transformers Are More Efficient Language ModelsCode1
AVocaDo: Strategy for Adapting Vocabulary to Downstream DomainCode1
Spanish Legalese Language Model and CorporaCode1
ClimateBert: A Pretrained Language Model for Climate-Related TextCode1
LMSOC: An Approach for Socially Sensitive PretrainingCode1
Training Deep Neural Networks with Adaptive Momentum Inspired by the Quadratic OptimizationCode1
GNN-LM: Language Modeling based on Global Contexts via GNNCode1
An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language ModelsCode1
Improving Transformers with Probabilistic Attention KeysCode1
Hydra: A System for Large Multi-Model Deep LearningCode1
Invariant Language ModelingCode1
A Good Prompt Is Worth Millions of Parameters: Low-resource Prompt-based Learning for Vision-Language ModelsCode1
Coherence boosting: When your pretrained language model is not paying enough attentionCode1
Show:102550
← PrevPage 53 of 284Next →

No leaderboard results yet.