Skip to main content

Efficient Dataset Distillation via Minimax Diffusion

Publication ,  Conference
Gu, J; Vahidian, S; Kungurtsev, V; Wang, H; Jiang, W; You, Y; Chen, Y
Published in: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
January 1, 2024

Dataset distillation reduces the storage and computational consumption of training a network by generating a small surrogate dataset that encapsulates rich information of the original large-scale one. However, previous distillation methods heavily rely on the sample-wise iterative optimization scheme. As the images-per-class (IPC) setting or image resolution grows larger, the necessary computation will demand overwhelming time and resources. In this work, we intend to incorporate generative diffusion techniques for computing the surrogate dataset. Observing that key factors for constructing an effective surrogate dataset are representativeness and diversity, we design additional minimax criteria in the generative training to enhance these facets for the generated images of diffusion models. We present a theoretical model of the process as hierarchical diffusion control demonstrating the flexibility of the diffusion process to target these criteria without jeopardizing the faithfulness of the sample to the desired distribution. The proposed method achieves state-of-the-art validation performance while demanding much less computational resources. Under the 100-IPC setting on Image Woof, our method requires less than one-twentieth the distillation time of previous methods, yet yields even better performance. Source code and generated data are available in https://github.com/vimar-gu/MinimaxDiffusion.

Duke Scholars

Published In

Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition

DOI

ISSN

1063-6919

Publication Date

January 1, 2024

Start / End Page

15793 / 15803
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Gu, J., Vahidian, S., Kungurtsev, V., Wang, H., Jiang, W., You, Y., & Chen, Y. (2024). Efficient Dataset Distillation via Minimax Diffusion. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (pp. 15793–15803). https://doi.org/10.1109/CVPR52733.2024.01495
Gu, J., S. Vahidian, V. Kungurtsev, H. Wang, W. Jiang, Y. You, and Y. Chen. “Efficient Dataset Distillation via Minimax Diffusion.” In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 15793–803, 2024. https://doi.org/10.1109/CVPR52733.2024.01495.
Gu J, Vahidian S, Kungurtsev V, Wang H, Jiang W, You Y, et al. Efficient Dataset Distillation via Minimax Diffusion. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 2024. p. 15793–803.
Gu, J., et al. “Efficient Dataset Distillation via Minimax Diffusion.” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2024, pp. 15793–803. Scopus, doi:10.1109/CVPR52733.2024.01495.
Gu J, Vahidian S, Kungurtsev V, Wang H, Jiang W, You Y, Chen Y. Efficient Dataset Distillation via Minimax Diffusion. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 2024. p. 15793–15803.

Published In

Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition

DOI

ISSN

1063-6919

Publication Date

January 1, 2024

Start / End Page

15793 / 15803