Keywords: dataset distillation
Abstract: Dataset Distillation aims to compress a large dataset into a significantly more compact, synthetic one without compromising the performance of the trained models.
To achieve this, existing methods use the agent model to extract information from the target dataset and embed it into the distilled dataset.
Consequently, the quality of extracted and embedded information determines the quality of the distilled dataset.
In this work, we find that existing methods introduce misaligned information in both information extraction and embedding stages.
To alleviate this, we propose Prioritize Alignment in Dataset Distillation (\textbf{PAD}), which aligns information from the following two perspectives.
1) We prune the target dataset according to the compressing ratio to filter the information that can be extracted by the agent model.
2) We use only deep layers of the agent model to perform the distillation to avoid excessively introducing low-level information.
This simple strategy effectively filters out misaligned information and brings non-trivial improvement for mainstream matching-based distillation algorithms.
Furthermore, built on trajectory matching, \textbf{PAD} achieves remarkable improvements on various benchmarks, achieving state-of-the-art performance.
Primary Area: applications to computer vision, audio, language, and other modalities
Code Of Ethics: I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics.
Submission Guidelines: I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide.
Anonymous Url: I certify that there is no URL (e.g., github page) that could be used to find authors’ identity.
No Acknowledgement Section: I certify that there is no acknowledgement section in this submission for double blind review.
Submission Number: 4897
Loading