Data-free Knowledge Distillation for Fine-grained Visual Categorization (2404.12037v1)
Abstract: Data-free knowledge distillation (DFKD) is a promising approach for addressing issues related to model compression, security privacy, and transmission restrictions. Although the existing methods exploiting DFKD have achieved inspiring achievements in coarse-grained classification, in practical applications involving fine-grained classification tasks that require more detailed distinctions between similar categories, sub-optimal results are obtained. To address this issue, we propose an approach called DFKD-FGVC that extends DFKD to fine-grained visual categorization~(FGVC) tasks. Our approach utilizes an adversarial distillation framework with attention generator, mixed high-order attention distillation, and semantic feature contrast learning. Specifically, we introduce a spatial-wise attention mechanism to the generator to synthesize fine-grained images with more details of discriminative parts. We also utilize the mixed high-order attention mechanism to capture complex interactions among parts and the subtle differences among discriminative features of the fine-grained categories, paying attention to both local features and semantic context relationships. Moreover, we leverage the teacher and student models of the distillation framework to contrast high-level semantic feature maps in the hyperspace, comparing variances of different categories. We evaluate our approach on three widely-used FGVC benchmarks (Aircraft, Cars196, and CUB200) and demonstrate its superior performance.
- Robust optimization, volume 28. Princeton university press, 2009.
- Poof: Part-based one-vs.-one features for fine-grained categorization, face verification, and attribute estimation. In CVPR, pages 955–962, 2013.
- Higher-order integration of hierarchical convolutional activations for fine-grained visual categorization. In ICCV, pages 511–520, 2017.
- Mixed high-order attention network for person re-identification. In CVPR, pages 371–381, 2019.
- Data-free learning of student networks. In ICCV, pages 3514–3522, 2019.
- Wasserstein contrastive representation distillation. In CVPR, pages 16296–16305, 2021.
- A simple framework for contrastive learning of visual representations. In ICML, pages 1597–1607. PMLR, 2020.
- Data-free network quantization with adversarial knowledge distillation. In CVPR, pages 710–711, 2020.
- Selective sparse sampling for fine-grained image recognition. In ICCV, pages 6599–6608, 2019.
- Momentum adversarial distillation: Handling large distribution shifts in data-free knowledge distillation. In Advances in NeurIPS, 2022.
- Data-free adversarial distillation. arXiv preprint arXiv:1912.11006, 2019.
- Contrastive model inversion for data-free knowledge distillation. In IJCAI, 2021.
- Weakly supervised complementary parts models for fine-grained image classification from the bottom up. In CVPR, pages 3034–3043, 2019.
- Attention mechanisms in computer vision: A survey. Computational Visual Media, 8(3):331–368, 2022.
- Deep residual learning for image recognition. In CVPR, pages 770–778, 2016.
- Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
- Squeeze-and-excitation networks. pages 7132–7141, 2018.
- Part-stacked cnn for fine-grained visual categorization. In CVPR, pages 1173–1182, 2016.
- Attention convolutional binary neural tree for fine-grained visual categorization. In CVPR, pages 10468–10477, 2020.
- An attention-driven two-stage clustering method for unsupervised person re-identification. In ECCV, pages 20–36. Springer, 2020.
- Supervised contrastive learning. Advances in NeurIPS, 33:18661–18673, 2020.
- Adam: A method for stochastic optimization. In ICLR, 2014.
- 3d object representations for fine-grained categorization. In ICCV workshops, pages 554–561, 2013.
- Imagenet classification with deep convolutional neural networks. Advances in NeurIPS, 25, 2012.
- Pruning filters for efficient convnets. In ICLR, 2016.
- Filtration and distillation: Enhancing region attention for fine-grained visual categorization. In AAAI, volume 34, pages 11555–11562, 2020.
- Zero-shot adversarial quantization. In CVPR, pages 1512–1521, 2021.
- Data-free knowledge transfer: A survey. arXiv preprint arXiv:2112.15278, 2021.
- Fine-grained visual classification of aircraft. arXiv preprint arXiv:1306.5151, 2013.
- Zero-shot knowledge transfer via adversarial belief matching. Advances in NeurIPS, 32, 2019.
- Multi-objective matrix normalization for fine-grained visual recognition. IEEE Transactions on Image Processing, 29:4996–5009, 2020.
- Spectral normalization for generative adversarial networks. In ICLR, 2018.
- Zero-shot knowledge distillation in deep networks. In ICML, pages 4743–4751. PMLR, 2019.
- Bam: Bottleneck attention module. BMVC, 2018.
- Unsupervised representation learning with deep convolutional generative adversarial networks. In ICLR, 2015.
- Mobilenetv2: Inverted residuals and linear bottlenecks. In CVPR, pages 4510–4520, 2018.
- Channel-wise knowledge distillation for dense prediction. In ICCV, pages 5311–5320, 2021.
- Deep convolutional generative adversarial network. In Artificial Intelligence with Python, pages 289–301. Springer, 2022.
- Contrastive representation distillation. arXiv preprint arXiv:1910.10699, 2019.
- Residual attention network for image classification. In CVPR, pages 3156–3164, 2017.
- Graph-propagation based correlation learning for weakly supervised fine-grained image classification. In AAAI, volume 34, pages 12289–12296, 2020.
- Fine-grained image analysis with deep learning: A survey. TPAMI, 2021.
- Caltech-ucsd birds 200. 2010.
- Cbam: Convolutional block attention module. In ECCV, pages 3–19, 2018.
- Attention deep model with multi-scale deep supervision for person re-identification. IEEE Trans. ETCI, 5(1):70–78, 2021.
- Attention-aware compositional network for person re-identification. In CVPR, pages 2119–2128, 2018.
- Dreaming to distill: Data-free knowledge transfer via deepinversion. In CVPR, pages 8715–8724, 2020.
- Part-based r-cnns for fine-grained category detection. In ECCV, pages 834–849. Springer, 2014.
- Improving neural network quantization without retraining using outlier channel splitting. In ICML, pages 7543–7552. PMLR, 2019.
- Graph-based high-order relation discovery for fine-grained recognition. In CVPR, pages 15079–15088, 2021.
- Learning multi-attention convolutional neural network for fine-grained image recognition. In ICCV, pages 5209–5217, 2017.
- Channel distillation: Channel-wise attention for knowledge distillation. arXiv preprint arXiv:2006.01683, 2020.
- Learning attentive pairwise interaction for fine-grained classification. In AAAI, volume 34, pages 13130–13137, 2020.