Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
134 tokens/sec
GPT-4o
10 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Improve Cross-Architecture Generalization on Dataset Distillation (2402.13007v1)

Published 20 Feb 2024 in cs.LG and cs.CV

Abstract: Dataset distillation, a pragmatic approach in machine learning, aims to create a smaller synthetic dataset from a larger existing dataset. However, existing distillation methods primarily adopt a model-based paradigm, where the synthetic dataset inherits model-specific biases, limiting its generalizability to alternative models. In response to this constraint, we propose a novel methodology termed "model pool". This approach involves selecting models from a diverse model pool based on a specific probability distribution during the data distillation process. Additionally, we integrate our model pool with the established knowledge distillation approach and apply knowledge distillation to the test process of the distilled dataset. Our experimental results validate the effectiveness of the model pool approach across a range of existing models while testing, demonstrating superior performance compared to existing methodologies.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. L. Breiman. Random forests. Machine Learning, 45:5–32, 2001.
  2. Dataset distillation by matching training trajectories, 2022.
  3. Selection via proxy: Efficient data selection for deep learning. arXiv preprint arXiv:1906.11829, 2019.
  4. Dc-bench: Dataset condensation benchmark, 2022.
  5. Scaling up dataset distillation to imagenet-1k with constant memory, 2023.
  6. Remember the past: Distilling datasets into addressable memories for neural networks, 2022.
  7. Minimizing the accumulated trajectory error to improve dataset distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 3749–3758, 2023.
  8. Sequential subset matching for dataset distillation, 2023.
  9. Model-agnostic meta-learning for fast adaptation of deep networks. In International conference on machine learning, pages 1126–1135. PMLR, 2017.
  10. A desicion-theoretic generalization of on-line learning and an application to boosting. In Paul Vitányi, editor, Computational Learning Theory, pages 23–37, Berlin, Heidelberg, 1995. Springer Berlin Heidelberg.
  11. Knowledge distillation: A survey. International Journal of Computer Vision, 129(6):1789–1819, March 2021.
  12. Distilling the knowledge in a neural network, 2015.
  13. Data distribution search to select core-set for machine learning. In The 9th International Conference on Smart Media and Applications, pages 172–176, 2020.
  14. On divergence measures for bayesian pseudocoresets. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), 2022.
  15. Dataset condensation via efficient synthetic-data parameterization. In International Conference on Machine Learning, pages 11102–11118. PMLR, 2022.
  16. Dataset condensation with contrastive signals, 2022.
  17. Knowledge condensation distillation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 19–35, 2022.
  18. Cat: Balanced continual graph learning with graph condensation. In Proceedings of the IEEE International Conference on Data Mining (ICDM), 2023.
  19. Reducing catastrophic forgetting with learning on synthetic data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Workshop, pages 4321–4326, 2020.
  20. Data distillation: A survey. arXiv preprint arXiv:2301.04272, 2023.
  21. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research, 15(1):1929–1958, 2014.
  22. Does knowledge distillation really work?, 2021.
  23. SecDD: Efficient and secure method for remotely training neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), pages 15897–15898, 2021.
  24. Can we achieve robustness from data alone? In Proceedings of the International Conference on Machine Learning (ICML), Workshop, 2022.
  25. Dataset distillation. arXiv preprint arXiv:1811.10959, 2018.
  26. Condensed composite memory continual learning. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), pages 1–8, 2021.
  27. FedDM: Iterative distribution matching for communication-efficient federated learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 16323–16332, 2023.
  28. Distill gold from massive ores: Efficient dataset distillation via critical samples selection. arXiv preprint arXiv:2305.18381, 2023.
  29. Squeeze, recover and relabel: Dataset condensation at imagenet scale from a new perspective. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), 2023.
  30. Bo Zhao and Hakan Bilen. Dataset condensation with differentiable siamese augmentation, 2021.
  31. Bo Zhao and Hakan Bilen. Dataset condensation with differentiable siamese augmentation. In Proceedings of the International Conference on Machine Learning (ICML), pages 12674–12685, 2021.
  32. Bo Zhao and Hakan Bilen. Dataset condensation with distribution matching, 2022.
  33. Bo Zhao and Hakan Bilen. Synthesizing informative training samples with gan, 2022.
  34. Dataset condensation with gradient matching. arXiv preprint arXiv:2006.05929, 2020.
  35. Towards mitigating architecture overfitting in dataset distillation, 2023.
  36. Towards mitigating architecture overfitting in dataset distillation. arXiv preprint arXiv:2309.04195, 2023.
  37. Dataset distillation using neural feature regression. Advances in Neural Information Processing Systems, 35:9813–9827, 2022.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets