Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

CoroNetGAN: Controlled Pruning of GANs via Hypernetworks (2403.08261v1)

Published 13 Mar 2024 in cs.CV, cs.AI, and eess.IV

Abstract: Generative Adversarial Networks (GANs) have proven to exhibit remarkable performance and are widely used across many generative computer vision applications. However, the unprecedented demand for the deployment of GANs on resource-constrained edge devices still poses a challenge due to huge number of parameters involved in the generation process. This has led to focused attention on the area of compressing GANs. Most of the existing works use knowledge distillation with the overhead of teacher dependency. Moreover, there is no ability to control the degree of compression in these methods. Hence, we propose CoroNet-GAN for compressing GAN using the combined strength of differentiable pruning method via hypernetworks. The proposed method provides the advantage of performing controllable compression while training along with reducing training time by a substantial factor. Experiments have been done on various conditional GAN architectures (Pix2Pix and CycleGAN) to signify the effectiveness of our approach on multiple benchmark datasets such as Edges-to-Shoes, Horse-to-Zebra and Summer-to-Winter. The results obtained illustrate that our approach succeeds to outperform the baselines on Zebra-to-Horse and Summer-to-Winter achieving the best FID score of 32.3 and 72.3 respectively, yielding high-fidelity images across all the datasets. Additionally, our approach also outperforms the state-of-the-art methods in achieving better inference time on various smart-phone chipsets and data-types making it a feasible solution for deployment on edge devices.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Generative adversarial nets. In NIPS, 2014.
  2. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  3. Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60:84 – 90, 2012.
  4. Very deep convolutional networks for large-scale image recognition. CoRR, abs/1409.1556, 2015.
  5. Quantization and training of neural networks for efficient integer-arithmetic-only inference. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2704–2713, 2018.
  6. Pruning filters for efficient convnets. ArXiv, abs/1608.08710, 2017.
  7. Localization-aware channel pruning for object detection. Neurocomputing, 403:400–408, 2020.
  8. A systematic dnn weight pruning framework using alternating direction method of multipliers. In Proceedings of the European Conference on Computer Vision (ECCV), pages 184–199, 2018.
  9. Fitnets: Hints for thin deep nets. arXiv preprint arXiv:1412.6550, 2014.
  10. Learning efficient object detection models with knowledge distillation. Advances in neural information processing systems, 30, 2017.
  11. Compressing gans using knowledge distillation. ArXiv, abs/1902.00159, 2019.
  12. Distilling portable generative adversarial networks for image translation. In AAAI, 2020.
  13. Gan slimming: All-in-one gan compression by a unified optimization framework. ArXiv, abs/2008.11062, 2020.
  14. Content-aware gan compression. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 12151–12161, 2021.
  15. Online multi-granularity distillation for gan compression. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pages 6773–6783, 2021.
  16. Image-to-image translation with conditional adversarial networks. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5967–5976, 2017.
  17. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8798–8807, 2018.
  18. Deblurgan: Blind motion deblurring using conditional adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8183–8192, 2018.
  19. Image-to-image translation via hierarchical style disentanglement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8639–8648, 2021.
  20. Smoothing the disentangled latent style space for unsupervised image-to-image translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 10785–10794, June 2021.
  21. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European conference on computer vision (ECCV) workshops, pages 0–0, 2018.
  22. Attngan: Fine-grained text to image generation with attentional generative adversarial networks. 2018.
  23. Unpaired image-to-image translation using cycle-consistent adversarial networks. 2017 IEEE International Conference on Computer Vision (ICCV), pages 2242–2251, 2017.
  24. Qgan: Quantized generative adversarial networks. ArXiv, abs/1901.08263, 2019.
  25. Tinygan: Distilling biggan for conditional image generation. In ACCV, 2020.
  26. Slimmable generative adversarial networks. In AAAI, 2021.
  27. Wavelet knowledge distillation: Towards efficient image-to-image translation. ArXiv, abs/2203.06321, 2022.
  28. Region-aware knowledge distillation for efficient image-to-image translation. ArXiv, abs/2205.12451, 2022.
  29. Teachers do more than teach: Compressing image-to-image models. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 13595–13606, 2021.
  30. Dynamic filter networks. CoRR, abs/1605.09673, 2016.
  31. Hypernetwork Functional Image Representation, pages 496–510. 09 2019.
  32. Bayesian hypernetworks. 10 2017.
  33. Hypernetworks. 09 2016.
  34. Christoph von der Malsburg. The correlation theory of brain function. Models Neural Netw., 2, 01 1994.
  35. Jürgen Schmidhuber. Learning to Control Fast-Weight Memories: An Alternative to Dynamic Recurrent Networks. Neural Computation, 4(1):131–139, 01 1992.
  36. Multiple object recognition with visual attention. 12 2014.
  37. Fully-convolutional siamese networks for object tracking. CoRR, abs/1606.09549, 2016.
  38. Spatial transformer networks. CoRR, abs/1506.02025, 2015.
  39. Learning deconvolution network for semantic segmentation. CoRR, abs/1505.04366, 2015.
  40. Dhp: Differentiable meta pruning via hypernetworks. In Proceedings of the European Conference on Computer Vision, 2020.
  41. Unsupervised representation learning with deep convolutional generative adversarial networks. CoRR, abs/1511.06434, 2016.
  42. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
  43. Learning multiple layers of features from tiny images. 2009.
  44. Gan compression: Efficient architectures for interactive conditional gans. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5283–5293, 2020.
  45. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017.
  46. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2818–2826, 2016.
  47. Learning efficient gans for image translation via differentiable masks and co-attention distillation. IEEE Transactions on Multimedia, 2022.
  48. Revisiting discriminator in gan compression: A generator-discriminator cooperative compression scheme. In NeurIPS, 2021.
  49. Autogan-distiller: Searching to compress generative adversarial networks. In ICML, 2020.
  50. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434, 2015.
Citations (1)

Summary

We haven't generated a summary for this paper yet.