SOTAVerified

Memorization

Papers

Showing 151200 of 1088 papers

TitleStatusHype
Do Language Models Plagiarize?Code1
Evolving Decomposed Plasticity Rules for Information-Bottlenecked Meta-LearningCode1
ByGPT5: End-to-End Style-conditioned Poetry Generation with Token-free Language ModelsCode1
LoTUS: Large-Scale Machine Unlearning with a Taste of UncertaintyCode1
Do We Need Zero Training Loss After Achieving Zero Training Error?Code1
MemBench: Memorized Image Trigger Prompt Dataset for Diffusion ModelsCode1
Artificial Neural Variability for Deep Learning: On Overfitting, Noise Memorization, and Catastrophic ForgettingCode1
CodeJudge-Eval: Can Large Language Models be Good Judges in Code Understanding?Code1
A Preference-aware Meta-optimization Framework for Personalized Vehicle Energy Consumption EstimationCode1
Jointly Non-Sampling Learning for Knowledge Graph Enhanced RecommendationCode1
Early-Learning Regularization Prevents Memorization of Noisy LabelsCode1
Evidentiality-guided Generation for Knowledge-Intensive NLP TasksCode1
AlleNoise: large-scale text classification benchmark dataset with real-world label noiseCode1
Elephants Never Forget: Memorization and Learning of Tabular Data in Large Language ModelsCode1
Memorization for Good: Encryption with Autoregressive Language ModelsCode1
Memorization in NLP Fine-tuning MethodsCode1
Enhancing human learning via spaced repetition optimizationCode1
Ethicist: Targeted Training Data Extraction Through Loss Smoothed Soft Prompting and Calibrated Confidence EstimationCode1
Erasing Undesirable Influence in Diffusion ModelsCode1
Mitigating Unintended Memorization with LoRA in Federated Learning for LLMsCode1
Euler State Networks: Non-dissipative Reservoir ComputingCode1
Consensual Collaborative Training And Knowledge Distillation Based Facial Expression Recognition Under Noisy AnnotationsCode1
Data Contamination Quiz: A Tool to Detect and Estimate Contamination in Large Language ModelsCode1
SoK: Membership Inference Attacks on LLMs are Rushing Nowhere (and How to Fix It)Code1
Hierarchical Deconstruction of LLM Reasoning: A Graph-Based Framework for Analyzing Knowledge UtilizationCode1
Exploring Memorization in Adversarial TrainingCode1
Execution-Based Evaluation for Open-Domain Code GenerationCode1
Continual Memorization of Factoids in Large Language ModelsCode1
Knowledge-Augmented Reasoning Distillation for Small Language Models in Knowledge-Intensive TasksCode1
DASH: Warm-Starting Neural Network Training in Stationary Settings without Loss of PlasticityCode1
Continual Variational Autoencoder Learning via Online Cooperative MemorizationCode1
Data Contamination Can Cross Language BarriersCode1
Contrastive Learning with Boosted MemorizationCode1
Contrast to Divide: Self-Supervised Pre-Training for Learning with Noisy LabelsCode1
State-of-the-Art Augmented NLP Transformer models for direct and single-step retrosynthesisCode1
Federated Learning from Vision-Language Foundation Models: Theoretical Analysis and MethodCode1
Brain tumor segmentation using synthetic MR images -- A comparison of GANs and diffusion modelsCode1
A Unified Framework for Model EditingCode1
Pre-training Large Memory Language Models with Internal and External KnowledgeCode1
Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy LabelsCode1
Do PLMs Know and Understand Ontological Knowledge?Code1
Mitigating Memorization of Noisy Labels via Regularization between RepresentationsCode1
In-context Autoencoder for Context Compression in a Large Language ModelCode1
Large Language Models Are Not Strong Abstract ReasonersCode1
Learning to Generate Novel Scene Compositions from Single Images and VideosCode1
Generalization or Memorization: Data Contamination and Trustworthy Evaluation for Large Language ModelsCode1
Generalization in diffusion models arises from geometry-adaptive harmonic representationsCode1
C-SFDA: A Curriculum Learning Aided Self-Training Framework for Efficient Source Free Domain AdaptationCode1
Adaptive Early-Learning Correction for Segmentation from Noisy AnnotationsCode1
MemoNet: Memorizing All Cross Features' Representations Efficiently via Multi-Hash Codebook Network for CTR PredictionCode1
Show:102550
← PrevPage 4 of 22Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1PaLM-540B (few-shot, k=5)Accuracy95.4Unverified
2Gopher-280B (few-shot, k=5)Accuracy80Unverified
3PaLM-62B (few-shot, k=5)Accuracy77.7Unverified