SOTAVerified

Language Modelling

A language model is a model of natural language. Language models are useful for a variety of tasks, including speech recognition, machine translation, natural language generation (generating more human-like text), optical character recognition, route optimization, handwriting recognition, grammar induction, and information retrieval.

Large language models (LLMs), currently their most advanced form, are predominantly based on transformers trained on larger datasets (frequently using words scraped from the public internet). They have superseded recurrent neural network-based models, which had previously superseded the purely statistical models, such as word n-gram language model.

Source: Wikipedia

Papers

Showing 14011450 of 17610 papers

TitleStatusHype
GRITHopper: Decomposition-Free Multi-Hop Dense RetrievalCode1
V2Flow: Unifying Visual Tokenization and Large Language Model Vocabularies for Autoregressive Image GenerationCode1
VLScene: Vision-Language Guidance Distillation for Camera-Based 3D Semantic Scene CompletionCode1
Multi-Layer Visual Feature Fusion in Multimodal LLMs: Methods, Analysis, and Best PracticesCode1
L^2M: Mutual Information Scaling Law for Long-Context Language ModelingCode1
Words or Vision: Do Vision-Language Models Have Blind Faith in Text?Code1
InfiniSST: Simultaneous Translation of Unbounded Speech with Large Language ModelCode1
Superscopes: Amplifying Internal Feature Representations for Language Model InterpretationCode1
Enhancing Monocular 3D Scene Completion with Diffusion ModelCode1
Safety Tax: Safety Alignment Makes Your Large Reasoning Models Less ReasonableCode1
Protein Structure Tokenization: Benchmarking and New RecipeCode1
FANformer: Improving Large Language Models Through Effective Periodicity ModelingCode1
Towards General Visual-Linguistic Face Forgery Detection(V2)Code1
Emergent Symbolic Mechanisms Support Abstract Reasoning in Large Language ModelsCode1
Adaptive Attacks Break Defenses Against Indirect Prompt Injection Attacks on LLM AgentsCode1
Playing Pokémon Red via Deep Reinforcement LearningCode1
SeisMoLLM: Advancing Seismic Monitoring via Cross-modal Transfer with Pre-trained Large Language ModelCode1
Inverse Materials Design by Large Language Model-Assisted Generative FrameworkCode1
Steering Language Model to Stable Speech Emotion Recognition via Contextual Perception and Chain of ThoughtCode1
Language Model Fine-Tuning on Scaled Survey Data for Predicting Distributions of Public OpinionsCode1
ARS: Automatic Routing Solver with Large Language ModelsCode1
STeCa: Step-level Trajectory Calibration for LLM Agent LearningCode1
LabTOP: A Unified Model for Lab Test Outcome Prediction on Electronic Health RecordsCode1
CORBA: Contagious Recursive Blocking Attacks on Multi-Agent Systems Based on Large Language ModelsCode1
Is Safety Standard Same for Everyone? User-Specific Safety Evaluation of Large Language ModelsCode1
Collaborative Retrieval for Large Language Model-based Conversational Recommender SystemsCode1
Towards Text-Image Interleaved RetrievalCode1
G-Refer: Graph Retrieval-Augmented Large Language Model for Explainable RecommendationCode1
AdaSplash: Adaptive Sparse Flash AttentionCode1
Logic.py: Bridging the Gap between LLMs and Constraint SolversCode1
M-ABSA: A Multilingual Dataset for Aspect-Based Sentiment AnalysisCode1
Can LLM Watermarks Robustly Prevent Unauthorized Knowledge Distillation?Code1
APB: Accelerating Distributed Long-Context Inference by Passing Compressed Context Blocks across GPUsCode1
video-SALMONN-o1: Reasoning-enhanced Audio-visual Large Language ModelCode1
Market-Derived Financial Sentiment Analysis: Context-Aware Language Models for Crypto ForecastingCode1
DuplexMamba: Enhancing Real-time Speech Conversations with Duplex and Streaming CapabilitiesCode1
MMUnlearner: Reformulating Multimodal Machine Unlearning in the Era of Multimodal Large Language ModelsCode1
Knowledge Graph-Driven Retrieval-Augmented Generation: Integrating Deepseek-R1 with Weaviate for Advanced Chatbot ApplicationsCode1
Exposing Numeracy Gaps: A Benchmark to Evaluate Fundamental Numerical Abilities in Large Language ModelsCode1
Reading Your Heart: Learning ECG Words and Sentences via Pre-training ECG Language ModelCode1
Can Large Language Model Agents Balance Energy Systems?Code1
SelfElicit: Your Language Model Secretly Knows Where is the Relevant EvidenceCode1
JamendoMaxCaps: A Large Scale Music-caption Dataset with Imputed MetadataCode1
Small Language Model Makes an Effective Long Text ExtractorCode1
MGPATH: Vision-Language Model with Multi-Granular Prompt Learning for Few-Shot WSI ClassificationCode1
RALLRec: Improving Retrieval Augmented Large Language Model Recommendation with Representation LearningCode1
Implicit Language Models are RNNs: Balancing Parallelization and ExpressivityCode1
Jakiro: Boosting Speculative Decoding with Decoupled Multi-Head via MoECode1
DexVLA: Vision-Language Model with Plug-In Diffusion Expert for General Robot ControlCode1
UniCMs: A Unified Consistency Model For Efficient Multimodal Generation and UnderstandingCode1
Show:102550
← PrevPage 29 of 353Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Decay RNNValidation perplexity76.67Unverified
2GRUValidation perplexity53.78Unverified
3LSTMValidation perplexity52.73Unverified
4LSTMTest perplexity48.7Unverified
5Temporal CNNTest perplexity45.2Unverified
6TCNTest perplexity45.19Unverified
7GCNN-8Test perplexity44.9Unverified
8Neural cache model (size = 100)Test perplexity44.8Unverified
9Neural cache model (size = 2,000)Test perplexity40.8Unverified
10GPT-2 SmallTest perplexity37.5Unverified
#ModelMetricClaimedVerifiedStatus
1TCNTest perplexity108.47Unverified
2Seq-U-NetTest perplexity107.95Unverified
3GRU (Bai et al., 2018)Test perplexity92.48Unverified
4R-TransformerTest perplexity84.38Unverified
5Zaremba et al. (2014) - LSTM (medium)Test perplexity82.7Unverified
6Gal & Ghahramani (2016) - Variational LSTM (medium)Test perplexity79.7Unverified
7LSTM (Bai et al., 2018)Test perplexity78.93Unverified
8Zaremba et al. (2014) - LSTM (large)Test perplexity78.4Unverified
9Gal & Ghahramani (2016) - Variational LSTM (large)Test perplexity75.2Unverified
10Inan et al. (2016) - Variational RHNTest perplexity66Unverified
#ModelMetricClaimedVerifiedStatus
1LSTM (7 layers)Bit per Character (BPC)1.67Unverified
2HypernetworksBit per Character (BPC)1.34Unverified
3SHA-LSTM (4 layers, h=1024, no attention head)Bit per Character (BPC)1.33Unverified
4LN HM-LSTMBit per Character (BPC)1.32Unverified
5ByteNetBit per Character (BPC)1.31Unverified
6Recurrent Highway NetworksBit per Character (BPC)1.27Unverified
7Large FS-LSTM-4Bit per Character (BPC)1.25Unverified
8Large mLSTMBit per Character (BPC)1.24Unverified
9AWD-LSTM (3 layers)Bit per Character (BPC)1.23Unverified
10Cluster-Former (#C=512)Bit per Character (BPC)1.22Unverified
#ModelMetricClaimedVerifiedStatus
1Smaller Transformer 126M (pre-trained)Test perplexity33Unverified
2OPT 125MTest perplexity32.26Unverified
3Larger Transformer 771M (pre-trained)Test perplexity28.1Unverified
4OPT 1.3BTest perplexity19.55Unverified
5GPT-Neo 125MTest perplexity17.83Unverified
6OPT 2.7BTest perplexity17.81Unverified
7Smaller Transformer 126M (fine-tuned)Test perplexity12Unverified
8GPT-Neo 1.3BTest perplexity11.46Unverified
9Transformer 125MTest perplexity10.7Unverified
10GPT-Neo 2.7BTest perplexity10.44Unverified