SOTAVerified

Dataset Distillation

Dataset distillation is the task of synthesizing a small dataset such that models trained on it achieve high performance on the original large dataset. A dataset distillation algorithm takes as input a large real dataset to be distilled (training set), and outputs a small synthetic distilled dataset, which is evaluated via testing models trained on this distilled dataset on a separate real dataset (validation/test set). A good small distilled dataset is not only useful in dataset understanding, but has various applications (e.g., continual learning, privacy, neural architecture search, etc.).

Papers

Showing 51100 of 216 papers

TitleStatusHype
A Label is Worth a Thousand Images in Dataset DistillationCode1
DiM: Distilling Dataset into Generative ModelCode1
Minimizing the Accumulated Trajectory Error to Improve Dataset DistillationCode1
Scaling Up Dataset Distillation to ImageNet-1K with Constant MemoryCode1
Towards Trustworthy Dataset DistillationCode1
Dataset Distillation via FactorizationCode1
D^4M: Dataset Distillation via Disentangled Diffusion ModelCode1
A Large-Scale Study on Video Action Dataset CondensationCode1
Dataset Distillation via Vision-Language Category PrototypeCode1
Dataset Distillation with Convexified Implicit GradientsCode1
Dataset DistillationCode1
Distilling Datasets Into Less Than One ImageCode1
DataDAM: Efficient Dataset Distillation with Attention MatchingCode1
Emphasizing Discriminative Features for Dataset Distillation in Complex ScenariosCode1
Dataset Factorization for CondensationCode1
Embarassingly Simple Dataset DistillationCode1
Dataset Distillation with Infinitely Wide Convolutional NetworksCode0
Behaviour DistillationCode0
BEARD: Benchmarking the Adversarial Robustness for Dataset DistillationCode0
Dataset Distillation via Knowledge Distillation: Towards Efficient Self-Supervised Pre-Training of Deep NetworksCode0
Dataset Distillation via Adversarial Prediction MatchingCode0
Dataset Distillation Using Parameter PruningCode0
Curriculum Coarse-to-Fine Selection for High-IPC Dataset DistillationCode0
Dataset Distillation using Neural Feature RegressionCode0
A Comprehensive Survey of Dataset DistillationCode0
Importance-Aware Adaptive Dataset DistillationCode0
CONCORD: Concept-Informed Diffusion for Dataset DistillationCode0
Hyperbolic Dataset DistillationCode0
BACON: Bayesian Optimal Condensation Framework for Dataset DistillationCode0
AST: Effective Dataset Distillation through Alignment with Smooth and High-Quality Expert TrajectoriesCode0
Going Beyond Feature Similarity: Effective Dataset Distillation based on Class-Aware Conditional Mutual InformationCode0
Image Distillation for Safe Data Sharing in HistopathologyCode0
Compressed Gastric Image Generation Based on Soft-Label Dataset Distillation for Medical Data SharingCode0
Dataset Distillation for Offline Reinforcement LearningCode0
ATOM: Attention Mixer for Efficient Dataset DistillationCode0
Diversity-Driven Synthesis: Enhancing Dataset Distillation through Directed Weight AdjustmentCode0
Dataset distillation for memorized data: Soft labels can leak held-out teacher knowledgeCode0
Dataset Distillation for Medical Dataset SharingCode0
Distributional Dataset Distillation with Subtask DecompositionCode0
Color-Oriented Redundancy Reduction in Dataset DistillationCode0
Generative Dataset Distillation: Balancing Global Structure and Local DetailsCode0
Does Training with Synthetic Data Truly Protect Privacy?Code0
Distill the Best, Ignore the Rest: Improving Dataset Distillation with Loss-Value-Based PruningCode0
Distilling Long-tailed DatasetsCode0
A Survey on Dataset Distillation: Approaches, Applications and Future DirectionsCode0
Understanding Reconstruction Attacks with the Neural Tangent Kernel and Dataset DistillationCode0
Few-Shot Dataset Distillation via Translative Pre-TrainingCode0
Distilled One-Shot Federated LearningCode0
Discovering Galaxy Features via Dataset DistillationCode0
Dataset Distillation by Automatic Training TrajectoriesCode0
Show:102550
← PrevPage 2 of 5Next →

No leaderboard results yet.