SOTAVerified

Memorization

Papers

Showing 601650 of 1088 papers

TitleStatusHype
The Positivity of the Neural Tangent Kernel0
The Reasoning-Memorization Interplay in Language Models Is Mediated by a Single Direction0
The Separation Capacity of Random Neural Networks0
Uncovering Memorization Effect in the Presence of Spurious Correlations0
The Sooner The Better: Investigating Structure of Early Winning Lottery Tickets0
The statistical thermodynamics of generative diffusion models: Phase transitions, symmetry breaking and critical instability0
The SWE-Bench Illusion: When State-of-the-Art LLMs Remember Instead of Reason0
The Unreasonable Effectiveness of the Class-reversed Sampling in Tail Sample Memorization0
The Vendiscope: An Algorithmic Microscope For Data Collections0
Thinking Tokens for Language Modeling0
Think or Remember? Detecting and Directing LLMs Towards Memorization or Generalization0
Three Factors Influencing Minima in SGD0
Through a Compressed Lens: Investigating the Impact of Quantization on LLM Explainability and Interpretability0
Time-Aware Language Models as Temporal Knowledge Bases0
To Each (Textual Sequence) Its Own: Improving Memorized-Data Unlearning in Large Language Models0
Too Big to Fool: Resisting Deception in Language Models0
Extracting Training Data from Unconditional Diffusion Models0
Towards Better Generalization in Open-Domain Question Answering by Mitigating Context Memorization0
Towards Differential Relational Privacy and its use in Question Answering0
Towards GAN Benchmarks Which Require Generalization0
Towards Harnessing Feature Embedding for Robust Learning with Noisy Labels0
Towards Memorization-Free Diffusion Models0
Towards Model-Size Agnostic, Compute-Free, Memorization-based Inference of Deep Learning0
Towards the Memorization Effect of Neural Networks in Adversarial Training0
Trade-offs in Data Memorization via Strong Data Processing Inequalities0
Training Data Attribution: Was Your Model Secretly Trained On Data Created By Mine?0
Training Data Extraction From Pre-trained Language Models: A Survey0
Training Data Protection with Compositional Diffusion Models0
Training Dynamic based data filtering may not work for NLP datasets0
Training Dynamics of Deep Network Linear Regions0
Training Foundation Models as Data Compression: On Information, Model Weights and Copyright Law0
Training Large ASR Encoders with Differential Privacy0
Training Production Language Models without Memorizing User Data0
Training Text-to-Text Transformers with Privacy Guarantees0
TReMu: Towards Neuro-Symbolic Temporal Reasoning for LLM-Agents with Memory in Multi-Session Dialogues0
Trustworthy Machine Learning via Memorization and the Granular Long-Tail: A Survey on Interactions, Tradeoffs, and Beyond0
Uncovering Conceptual Blindspots in Generative Image Models Using Sparse Autoencoders0
Uncovering Latent Memories: Assessing Data Leakage and Memorization Patterns in Frontier AI Models0
Underestimated Privacy Risks for Minority Populations in Large Language Model Unlearning0
Understanding Activation Patterns in Artificial Neural Networks by Exploring Stochastic Processes0
Modeling Neural Networks with Privacy Using Neural Stochastic Differential Equations0
Understanding and Mitigating Memorization in Diffusion Models for Tabular Data0
Understanding Fact Recall in Language Models: Why Two-Stage Training Encourages Memorization but Mixed Training Teaches Knowledge0
Understanding Feature Selection and Feature Memorization in Recurrent Neural Networks0
Understanding Generalization in Diffusion Models via Probability Flow Distance0
Understanding Learning through the Lens of Dynamical Invariants0
Understanding Memorization in Generative Models via Sharpness in Probability Landscapes0
Understanding the Effect of the Long Tail on Neural Network Compression0
Understanding the Local Geometry of Generative Model Manifolds0
Understanding (Un)Intended Memorization in Text-to-Image Generative Models0
Show:102550
← PrevPage 13 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1PaLM-540B (few-shot, k=5)Accuracy95.4Unverified
2Gopher-280B (few-shot, k=5)Accuracy80Unverified
3PaLM-62B (few-shot, k=5)Accuracy77.7Unverified