SOTAVerified

Memorization

Papers

Showing 101150 of 1088 papers

TitleStatusHype
Learning to Generate Gradients for Test-Time Adaptation via Test-Time Training LayersCode1
FineZip : Pushing the Limits of Large Language Models for Practical Lossless Text CompressionCode1
Continual Variational Autoencoder Learning via Online Cooperative MemorizationCode1
Benchmarking Chinese Commonsense Reasoning of LLMs: From Chinese-Specifics to Reasoning-Memorization CorrelationsCode1
Learning with Noisy Correspondence for Cross-modal MatchingCode1
Learning with Noisy Labels Revisited: A Study Using Real-World Human AnnotationsCode1
Execution-Based Evaluation for Open-Domain Code GenerationCode1
Towards Adversarial Evaluations for Inexact Machine UnlearningCode1
Exploring Memorization in Adversarial TrainingCode1
Beyond Memorization: The Challenge of Random Memory Access in Language ModelsCode1
Erasing Undesirable Influence in Diffusion ModelsCode1
Advancing Cross-domain Discriminability in Continual Learning of Vision-Language ModelsCode1
Are Large Pre-Trained Language Models Leaking Your Personal Information?Code1
An Empirical Study of Memorization in NLPCode1
MemBench: Memorized Image Trigger Prompt Dataset for Diffusion ModelsCode1
Membership Inference Attacks and Defenses in Neural Network PruningCode1
Copyright Traps for Large Language ModelsCode1
MemLLM: Finetuning LLMs to Use An Explicit Read-Write MemoryCode1
Memorization for Good: Encryption with Autoregressive Language ModelsCode1
Memorization in NLP Fine-tuning MethodsCode1
MEOW: MEMOry Supervised LLM Unlearning Via Inverted FactsCode1
Mitigating Memorization In Language ModelsCode1
Elephants Never Forget: Memorization and Learning of Tabular Data in Large Language ModelsCode1
Bot or Human? Detecting ChatGPT Imposters with A Single QuestionCode1
Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence EstimationCode1
Multi-Objective Interpolation Training for Robustness to Label NoiseCode1
Forget-Me-Not: Learning to Forget in Text-to-Image Diffusion ModelsCode1
Capabilities of GPT-4 on Medical Challenge ProblemsCode1
Do We Need Zero Training Loss After Achieving Zero Training Error?Code1
Driving Style Representation in Convolutional Recurrent Neural Network Model of Driver IdentificationCode1
Do Language Models Plagiarize?Code1
Can Forward Gradient Match Backpropagation?Code1
DISC: Learning From Noisy Labels via Dynamic Instance-Specific Selection and CorrectionCode1
Do SSL Models Have Déjà Vu? A Case of Unintended Memorization in Self-supervised LearningCode1
ByGPT5: End-to-End Style-conditioned Poetry Generation with Token-free Language ModelsCode1
A Preference-aware Meta-optimization Framework for Personalized Vehicle Energy Consumption EstimationCode1
Dissecting Generation Modes for Abstractive Summarization Models via Ablation and AttributionCode1
Can Language Models Follow Multiple Turns of Entangled Instructions?Code1
Can LLM Graph Reasoning Generalize beyond Pattern Memorization?Code1
Can Neural Network Memorization Be Localized?Code1
DAT: Training Deep Networks Robust To Label-Noise by Matching the Feature DistributionsCode1
Evolving Decomposed Plasticity Rules for Information-Bottlenecked Meta-LearningCode1
Data Unlearning in Diffusion ModelsCode1
Eicient Non-Sampling Factorization Machines for Optimal Context-Aware RecommendationCode1
Elephants Never Forget: Testing Language Models for Memorization of Tabular DataCode1
Enhancing human learning via spaced repetition optimizationCode1
CodeJudge-Eval: Can Large Language Models be Good Judges in Code Understanding?Code1
Euler State Networks: Non-dissipative Reservoir ComputingCode1
Evidentiality-guided Generation for Knowledge-Intensive NLP TasksCode1
Deciphering the Factors Influencing the Efficacy of Chain-of-Thought: Probability, Memorization, and Noisy ReasoningCode1
Show:102550
← PrevPage 3 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1PaLM-540B (few-shot, k=5)Accuracy95.4Unverified
2Gopher-280B (few-shot, k=5)Accuracy80Unverified
3PaLM-62B (few-shot, k=5)Accuracy77.7Unverified