SOTAVerified

Memorization

Papers

Showing 451500 of 1088 papers

TitleStatusHype
Retrieval Helps or Hurts? A Deeper Dive into the Efficacy of Retrieval Augmentation to Language ModelsCode0
Artifacts or Abduction: How Do LLMs Answer Multiple-Choice Questions Without the Question?Code0
Amplifying Training Data Exposure through Fine-Tuning with Pseudo-Labeled MembershipsCode0
Towards Uncovering How Large Language Model Works: An Explainability Perspective0
Neural Information Organizing and Processing -- Neural Machines0
Information Complexity of Stochastic Convex Optimization: Applications to Generalization and Memorization0
Copyright Traps for Large Language ModelsCode1
The Mirrored Influence Hypothesis: Efficient Data Influence Estimation by Harnessing Forward PassesCode1
Future Prediction Can be a Strong Evidence of Good History Representation in Partially Observable Environments0
Social Evolution of Published Text and The Emergence of Artificial Intelligence Through Large Language Models and The Problem of Toxicity and Bias0
Wasserstein proximal operators describe score-based generative models and resolve memorization0
Revisiting Early-Learning Regularization When Federated Learning Meets Noisy Labels0
Selective Forgetting: Advancing Machine Unlearning Techniques and Evaluation in Language Models0
Analyzing the Neural Tangent Kernel of Periodically Activated Coordinate Networks0
Amortized Planning with Large-Scale Transformers: A Case Study on ChessCode4
EMN: Brain-inspired Elastic Memory Network for Quick Domain Adaptive Feature Mapping0
Déjà Vu Memorization in Vision-Language Models0
Human-Centered Privacy Research in the Age of Large Language Models0
Unconditional Latent Diffusion Models Memorize Patient Imaging Data: Implications for Openly Sharing Synthetic DataCode0
Conserve-Update-Revise to Cure Generalization and Robustness Trade-off in Adversarial TrainingCode0
Expressive Power of ReLU and Step Networks under Floating-Point Operations0
Do LLMs Dream of Ontologies?Code0
Memorization in Self-Supervised Learning Improves Downstream GeneralizationCode0
Critical Data Size of Language Models from a Grokking Perspective0
Understanding Learning through the Lens of Dynamical Invariants0
Imputation with Inter-Series Information from Prototypes for Irregular Sampled Time SeriesCode0
Erasing Undesirable Influence in Diffusion ModelsCode1
A Good Score Does not Lead to A Good Generative ModelCode0
Learning with Structural Labels for Learning with Noisy Labels0
Infinite dSprites for Disentangled Continual Learning: Separating Memory Edits from GeneralizationCode0
Noisy Correspondence Learning with Self-Reinforcing Errors Mitigation0
RoleEval: A Bilingual Role Evaluation Benchmark for Large Language ModelsCode1
BloomVQA: Assessing Hierarchical Multi-modal Comprehension0
FedDiv: Collaborative Noise Filtering for Federated Learning with Noisy LabelsCode0
Social Learning: Towards Collaborative Learning with Large Language Models0
Lifted RDT based capacity analysis of the 1-hidden layer treelike sign perceptrons neural networks0
Negative Pre-aware for Noisy Cross-modal MatchingCode1
Preserving Privacy Through Dememorization: An Unlearning Technique For Mitigating Memorization Risks In Language ModelsCode0
SoK: Unintended Interactions among Machine Learning Defenses and RisksCode0
Understanding (Un)Intended Memorization in Text-to-Image Generative Models0
Memory Triggers: Unveiling Memorization in Text-To-Image Generative Models through Word-Level Duplication0
Scalable Extraction of Training Data from (Production) Language Models0
Continual Referring Expression Comprehension via Dual Modular MemorizationCode0
Positional Description Matters for Transformers Arithmetic0
CSGNN: Conquering Noisy Node labels via Dynamic Class-wise Selection0
Sparse Low-rank Adaptation of Pre-trained Language ModelsCode1
Negotiated Representations for Machine Mearning ApplicationCode0
On Retrieval Augmentation and the Limitations of Language Model Training0
Does Pre-trained Language Model Actually Infer Unseen Links in Knowledge Graph Completion?0
Fuse to Forget: Bias Reduction and Selective Memorization through Model FusionCode0
Show:102550
← PrevPage 10 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1PaLM-540B (few-shot, k=5)Accuracy95.4Unverified
2Gopher-280B (few-shot, k=5)Accuracy80Unverified
3PaLM-62B (few-shot, k=5)Accuracy77.7Unverified