SOTAVerified

Language Modeling

Papers

Showing 1135111400 of 14182 papers

TitleStatusHype
Cross-lingual Transfer Learning for Pre-trained Contextualized Language Models0
Domain-slot Relationship Modeling using a Pre-trained Language Encoder for Multi-Domain Dialogue State Tracking0
Context-Aware Temperature for Language Modeling0
BROS: A Pre-trained Language Model for Understanding Texts in Document0
Towards Practical Second Order Optimization for Deep Learning0
Transformer-QL: A Step Towards Making Transformer Network Quadratically Large0
Translation Memory Guided Neural Machine Translation0
Non-iterative Parallel Text Generation via Glancing Transformer0
Refine and Imitate: Reducing Repetition and Inconsistency in Dialogue Generation via Reinforcement Learning and Human Demonstration0
TaskSet: A Dataset of Optimization TasksCode0
Pretrain Knowledge-Aware Language Models0
Synthesizer: Rethinking Self-Attention for Transformer Models0
MONGOOSE: A Learnable LSH Framework for Efficient Neural Network Training0
Subformer: A Parameter Reduced Transformer0
SEQUENCE-LEVEL FEATURES: HOW GRU AND LSTM CELLS CAPTURE N-GRAMS0
Memory Representation in Transformer0
Adding Recurrence to Pretrained Transformers0
Discovering Autoregressive Orderings with Variational InferenceCode1
Block Skim Transformer for Efficient Question Answering0
K-PLUG: KNOWLEDGE-INJECTED PRE-TRAINED LANGUAGE MODEL FOR NATURAL LANGUAGE UNDERSTANDING AND GENERATIONCode1
Learning Chess Blindfolded0
Representation and Bias in Multilingual NLP: Insights from Controlled Experiments on Conditional Language Modeling0
On the use of linguistic similarities to improve Neural Machine Translation for African Languages0
Transformer protein language models are unsupervised structure learners0
ROMUL: Scale Adaptative Population Based Training0
Universal Sentence Representations Learning with Conditional Masked Language Model0
Syntactic Relevance XLNet Word Embedding Generation in Low-Resource Machine Translation0
Not All Memories are Created Equal: Learning to ExpireCode1
SCoRe: Pre-Training for Context Representation in Conversational Semantic Parsing0
The Pile: An 800GB Dataset of Diverse Text for Language ModelingCode2
Unified Mandarin TTS Front-end Based on Distilled BERT ModelCode1
XLM-T: Scaling up Multilingual Machine Translation with Pretrained Cross-lingual Transformer Encoders0
Studying Strategically: Learning to Mask for Closed-book QA0
Verb Knowledge Injection for Multilingual Event Processing0
Shortformer: Better Language Modeling using Shorter InputsCode1
AraGPT2: Pre-Trained Transformer for Arabic Language GenerationCode1
Directed Beam Search: Plug-and-Play Lexically Constrained Language GenerationCode0
ERNIE-Doc: A Retrospective Long-Document Modeling TransformerCode0
CoCoLM: COmplex COmmonsense Enhanced Language Model with Discourse RelationsCode0
AraELECTRA: Pre-Training Text Discriminators for Arabic Language UnderstandingCode1
Enhancing Pre-trained Language Model with Lexical Simplification0
Can Sequence-to-Sequence Models Crack Substitution Ciphers?0
SemGloVe: Semantic Co-occurrences for GloVe from BERTCode0
LayoutLMv2: Multi-modal Pre-training for Visually-Rich Document UnderstandingCode0
Generating Adversarial Examples in Chinese Texts Using Sentence-Pieces0
CMV-BERT: Contrastive multi-vocab pretraining of BERT0
Generating Query Focused Summaries from Query-Free ResourcesCode1
General Mechanism of Evolution Shared by Proteins and WordsCode0
Universal Sentence Representation Learning with Conditional Masked Language Model0
Assessment of the Relative Importance of different hyper-parameters of LSTM for an IDS0
Show:102550
← PrevPage 228 of 284Next →

No leaderboard results yet.