DD-Ranking: Rethinking the Evaluation of Dataset Distillation

📅 2025-05-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current data distillation (DD) evaluation heavily relies on test accuracy, making it susceptible to confounding factors such as soft labels and strong data augmentations—thus failing to reflect the intrinsic information quality of distilled images and compromising evaluation fidelity and reproducibility. To address this, we propose DD-Ranking: the first decoupled, robust, unified evaluation framework for DD. It isolates external technical influences via rank-consistency analysis, cross-model generalization testing, and ablation-driven benchmarking, focusing instead on the genuine information gain encoded in distilled data. Additionally, zero-shot transfer evaluation is introduced to assess the fundamental representational capacity of distilled samples. Experiments on CIFAR and ImageNet-1K reveal that performance gains reported for several state-of-the-art methods stem from “evaluation contamination” rather than true distillation efficacy. DD-Ranking substantially improves method discriminability, evaluation fairness, and result reproducibility.

Technology Category

Application Category

📝 Abstract
In recent years, dataset distillation has provided a reliable solution for data compression, where models trained on the resulting smaller synthetic datasets achieve performance comparable to those trained on the original datasets. To further improve the performance of synthetic datasets, various training pipelines and optimization objectives have been proposed, greatly advancing the field of dataset distillation. Recent decoupled dataset distillation methods introduce soft labels and stronger data augmentation during the post-evaluation phase and scale dataset distillation up to larger datasets (e.g., ImageNet-1K). However, this raises a question: Is accuracy still a reliable metric to fairly evaluate dataset distillation methods? Our empirical findings suggest that the performance improvements of these methods often stem from additional techniques rather than the inherent quality of the images themselves, with even randomly sampled images achieving superior results. Such misaligned evaluation settings severely hinder the development of DD. Therefore, we propose DD-Ranking, a unified evaluation framework, along with new general evaluation metrics to uncover the true performance improvements achieved by different methods. By refocusing on the actual information enhancement of distilled datasets, DD-Ranking provides a more comprehensive and fair evaluation standard for future research advancements.
Problem

Research questions and friction points this paper is trying to address.

Is accuracy a reliable metric for dataset distillation evaluation?
Do additional techniques overshadow synthetic image quality improvements?
Proposing DD-Ranking for fairer dataset distillation performance assessment.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposes DD-Ranking evaluation framework
Introduces new general evaluation metrics
Focuses on distilled datasets information enhancement
🔎 Similar Papers
No similar papers found.
Zekai Li
Zekai Li
PhD Student at UC San Diego
efficient deep learningdata-centric AI
Xinhao Zhong
Xinhao Zhong
Harbin Institute of Technology, Shenzhen
Data-centric AIEffiecient AI
Samir Khaki
Samir Khaki
University of Toronto
Artificial Intelligence
Zhiyuan Liang
Zhiyuan Liang
University of Science and Technology
Large Language ModelsMultimodal UnderstandingEfficient Learning
Y
Yuhao Zhou
Mingjia Shi
Mingjia Shi
Somewhere on the Earth
Learning TheoryData ScienceResource Preserving
Ziqiao Wang
Ziqiao Wang
Assistant Professor of Computer Science, Tongji University
Machine LearningStatistical Learning TheoryDeep LearningInformation Theory
Xuanlei Zhao
Xuanlei Zhao
National University of Singapore, CS PhD student
Efficient Deep LearningMachine Learning SystemHigh Performance Computing
Wangbo Zhao
Wangbo Zhao
National University of Singapore
Efficient Deep LearningDynamic Neural NetworkMultimodal Model
Z
Ziheng Qin
M
Mengxuan Wu
P
Pengfei Zhou
H
Haonan Wang
D
David Junhao Zhang
J
Jia-Wei Liu
Shaobo Wang
Shaobo Wang
Shanghai Jiao Tong University
Large Language ModelData-Centric AIData SynthesisData SelectionExplainable AI
Dai Liu
Dai Liu
Munich Technical University
Network CompressionDataset DistillationAI on HardwareAI and systems
Linfeng Zhang
Linfeng Zhang
DP Technology; AI for Science Institute
AI for Sciencemulti-scale modelingmolecular simulationdrug/materials design
Guang Li
Guang Li
Assistant Professor, Hokkaido University
Dataset DistillationSelf-Supervised LearningData-Centric AIMedical Image Analysis
K
Kun Wang
Z
Zheng Zhu
Z
Zhiheng Ma
Joey Tianyi Zhou
Joey Tianyi Zhou
A*STAR and NUS
Efficient AIRobust & Safe AI
Jiancheng Lv
Jiancheng Lv
University of Science and Technology of China
Operations ManagementMarketing
Y
Yaochu Jin
P
Peihao Wang
Kaipeng Zhang
Kaipeng Zhang
Shanghai AI Laboratory
LLMMultimodal LLMsAIGC
Lingjuan Lyu
Lingjuan Lyu
Sony
Foundation ModelsFederated LearningResponsible AI
Y
Yiran Huang
Zeynep Akata
Zeynep Akata
Professor at Technical University of Munich and Director at Helmholtz Munich
Machine LearningVision and LanguageZero-Shot Learning
Zhiwei Deng
Zhiwei Deng
Google DeepMind, Princeton University
Computer VisionMachine LearningDeep Learning
Xindi Wu
Xindi Wu
Princeton University
Computer VisionMachine Learning
George Cazenavette
George Cazenavette
MIT
artificial intelligencemachine learningcomputer vision
Yuzhang Shang
Yuzhang Shang
Assistant Professor at University of Central Florida
Efficient/Scalable AIDeep LearningMLCVNLP
Justin Cui
Justin Cui
PhD student at University of California, Los Angeles
Machine LearningComputer VisionArtifical Intelligence
Jindong Gu
Jindong Gu
Google Research & DeepMind, University of Oxford
Trustworthy AIAI SafetyMultimodal AI
Q
Qian Zheng
H
Hao Ye
S
Shuo Wang
Xiaobo Wang
Xiaobo Wang
University of Science and Technology of China
Natural Language Processing
Y
Yan Yan
Angela Yao
Angela Yao
National University of Singapore
computer visiondeep learningmachine learning
M
Mike Zheng Shou
Tianlong Chen
Tianlong Chen
Assistant Professor, CS@UNC Chapel Hill; Chief AI Scientist, hireEZ
Machine LearningAI4ScienceComputer VisionSparsity
Hakan Bilen
Hakan Bilen
University of Edinburgh
Computer VisionMachine Learning
Baharan Mirzasoleiman
Baharan Mirzasoleiman
UCLA
Machine LearningOptimizationSubmodularityML SustainabilityData-quality
M
M. Kellis
K
Konstantinos N. Plataniotis
Z
Zhangyang Wang
B
Bo Zhao
Yang You
Yang You
Postdoc, Stanford University
3D visioncomputer graphicscomputational geometry
K
Kai Wang