SOTAVerified

Memorization

Papers

Showing 201250 of 1088 papers

TitleStatusHype
Self-Attentive Associative MemoryCode1
Generalization through Memorization: Nearest Neighbor Language ModelsCode1
Lifelong Sequential Modeling with Personalized Memorization for User Response PredictionCode1
Enhancing human learning via spaced repetition optimizationCode1
Learning to Remember More with Less MemorizationCode1
Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy LabelsCode1
Memorization Precedes Generation: Learning Unsupervised GANs with Memory NetworksCode1
mixup: Beyond Empirical Risk MinimizationCode1
Wide & Deep Learning for Recommender SystemsCode1
What Should LLMs Forget? Quantifying Personal Data in LLMs for Right-to-Be-Forgotten Requests0
Entropy-Memorization Law: Evaluating Memorization Difficulty of Data in LLMs0
MMReason: An Open-Ended Multi-Modal Multi-Step Reasoning Benchmark for MLLMs Toward AGICode0
Listener-Rewarded Thinking in VLMs for Image Preferences0
Where to find Grokking in LLM Pretraining? Monitor Memorization-to-Generalization without Test0
Leaner Training, Lower Leakage: Revisiting Memorization in LLM Fine-Tuning with LoRA0
Counterfactual Influence as a Distributional Quantity0
Uncovering Conceptual Blindspots in Generative Image Models Using Sparse Autoencoders0
A Random Matrix Analysis of In-context Memorization for Nonlinear Attention0
Robots and Children that Learn Together : Improving Knowledge Retention by Teaching Peer-Like Interactive Robots0
In-Context Learning Strategies Emerge Rationally0
LexiMark: Robust Watermarking via Lexical Substitutions to Enhance Membership Verification of an LLM's Textual Training DataCode0
Winter Soldier: Backdooring Language Models at Pre-Training with Indirect Data Poisoning0
Dataset distillation for memorized data: Soft labels can leak held-out teacher knowledgeCode0
Capacity Matters: a Proof-of-Concept for Transformer Memorization on Real-World DataCode0
Less is More: Undertraining Experts Improves Model Upcycling0
Sharpness-Aware Machine Unlearning0
The SWE-Bench Illusion: When State-of-the-Art LLMs Remember Instead of Reason0
Restoring Gaussian Blurred Face Images for Deanonymization Attacks0
Diffusion models under low-noise regimeCode0
Private Memorization Editing: Turning Memorization into a Defense to Strengthen Data Privacy in Large Language ModelsCode0
SoK: Data Reconstruction Attacks Against Machine Learning Models: Definition, Metrics, and Benchmark0
Simple Yet Effective: Extracting Private Data Across Clients in Federated Fine-Tuning of Large Language Models0
Quantifying Cross-Modality Memorization in Vision-Language Models0
Membership Inference Attacks on Sequence Models0
Beyond Memorization: A Rigorous Evaluation Framework for Medical Knowledge EditingCode0
Trade-offs in Data Memorization via Strong Data Processing Inequalities0
How much do language models memorize?0
Bayesian Perspective on Memorization and Reconstruction0
Kernel-Smoothed Scores for Denoising Diffusion: A Bias-Variance Study0
Benchmarking Abstract and Reasoning Abilities Through A Theoretical PerspectiveCode0
OWL: Probing Cross-Lingual Recall of Memorized Texts via World LiteratureCode0
Navigating the Latent Space Dynamics of Neural Models0
Memorization to Generalization: Emergence of Diffusion Models from Associative Memory0
What is Adversarial Training for Diffusion Models?0
Understanding Generalization in Diffusion Models via Probability Flow Distance0
Grokking ExPLAIND: Unifying Model, Data, and Training Attribution to Study Model BehaviorCode0
Emergent LLM behaviors are observationally equivalent to data leakageCode0
Spurious Privacy Leakage in Neural Networks0
Querying Kernel Methods Suffices for Reconstructing their Training DataCode0
Discovering Forbidden Topics in Language Models0
Show:102550
← PrevPage 5 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1PaLM-540B (few-shot, k=5)Accuracy95.4Unverified
2Gopher-280B (few-shot, k=5)Accuracy80Unverified
3PaLM-62B (few-shot, k=5)Accuracy77.7Unverified