Dataset Condensation
26 papers with code • 0 benchmarks • 0 datasets
Condense the full dataset into a tiny set of synthetic data.
Benchmarks
These leaderboards are used to track progress in Dataset Condensation
Most implemented papers
You Only Condense Once: Two Rules for Pruning Condensed Datasets
However, these scenarios have two significant challenges: 1) the varying computational resources available on the devices require a dataset size different from the pre-defined condensed dataset, and 2) the limited computational resources often preclude the possibility of conducting additional condensation processes.
Generalized Large-Scale Data Condensation via Various Backbone and Statistical Matching
We call this perspective "generalized matching" and propose Generalized Various Backbone and Statistical Matching (G-VBSM) in this work, which aims to create a synthetic dataset with densities, ensuring consistency with the complete dataset across various backbones, layers, and statistics.
Dataset Condensation Driven Machine Unlearning
To achieve this goal, we propose new dataset condensation techniques and an innovative unlearning scheme that strikes a balance between machine unlearning privacy, utility, and efficiency.
Is Adversarial Training with Compressed Datasets Effective?
This synthetic dataset retains the essential information of the original dataset, enabling models trained on it to achieve performance levels comparable to those trained on the full dataset.
Multisize Dataset Condensation
These two challenges connect to the "subset degradation problem" in traditional dataset condensation: a subset from a larger condensed dataset is often unrepresentative compared to directly condensing the whole dataset to that smaller size.
Distilling Datasets Into Less Than One Image
Current methods frame this as maximizing the distilled classification accuracy for a budget of K distilled images-per-class, where K is a positive integer.