Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

EGAIN: Extended GAn INversion (2312.15116v1)

Published 22 Dec 2023 in cs.CV and cs.AI

Abstract: Generative Adversarial Networks (GANs) have witnessed significant advances in recent years, generating increasingly higher quality images, which are non-distinguishable from real ones. Recent GANs have proven to encode features in a disentangled latent space, enabling precise control over various semantic attributes of the generated facial images such as pose, illumination, or gender. GAN inversion, which is projecting images into the latent space of a GAN, opens the door for the manipulation of facial semantics of real face images. This is useful for numerous applications such as evaluating the performance of face recognition systems. In this work, EGAIN, an architecture for constructing GAN inversion models, is presented. This architecture explicitly addresses some of the shortcomings in previous GAN inversion models. A specific model with the same name, egain, based on this architecture is also proposed, demonstrating superior reconstruction quality over state-of-the-art models, and illustrating the validity of the EGAIN architecture.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. Rameen Abdal, Yipeng Qin and Peter Wonka “Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space?”, 2019 arXiv:1904.03189 [cs.CV]
  2. Yuval Alaluf, Or Patashnik and Daniel Cohen-Or “ReStyle: A Residual-Based StyleGAN Encoder via Iterative Refinement” In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021
  3. Lacey Best-Rowden and Anil K. Jain “Learning Face Image Quality From Human Assessments” In IEEE Transactions on Information Forensics and Security 13.12, 2018, pp. 3064–3077 DOI: 10.1109/TIFS.2018.2799585
  4. “Ensembling with Deep Generative Views”, 2021 arXiv:2104.14551 [cs.CV]
  5. “SSD-GAN: Measuring the Realness in the Spatial and Spectral Domains”, 2020 arXiv:2012.05535 [cs.CV]
  6. “ArcFace: Additive Angular Margin Loss for Deep Face Recognition”, 2019 arXiv:1801.07698 [cs.CV]
  7. Ricard Durall, Margret Keuper and Janis Keuper “Watch your Up-Convolution: CNN Based Generative Deep Neural Networks are Failing to Reproduce Spectral Distributions”, 2020 arXiv:2003.01826 [cs.CV]
  8. Tarik Dzanic, Karan Shah and Freddie Witherden “Fourier Spectrum Discrepancies in Deep Network Generated Images”, 2020 arXiv:1911.06465 [eess.IV]
  9. “SWAGAN: A Style-based Wavelet-driven Generative Model”, 2021 arXiv:2102.06108 [cs.CV]
  10. “Generative Adversarial Networks”, 2014 arXiv:1406.2661 [stat.ML]
  11. Marcel Grimmer, Raghavendra Ramachandra and Christoph Busch “Deep Face Age Progression: A Survey” In IEEE Access 9, 2021, pp. 83376–83393 DOI: 10.1109/ACCESS.2021.3085835
  12. “Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization”, 2017 arXiv:1703.06868 [cs.CV]
  13. Tero Karras, Samuli Laine and Timo Aila “A Style-Based Generator Architecture for Generative Adversarial Networks”, 2019 arXiv:1812.04948 [cs.NE]
  14. “Alias-Free Generative Adversarial Networks”, 2021 arXiv:2106.12423 [cs.CV]
  15. “Analyzing and Improving the Image Quality of StyleGAN”, 2020 arXiv:1912.04958 [cs.CV]
  16. “Progressive Growing of GANs for Improved Quality, Stability, and Variation”, 2018 arXiv:1710.10196 [cs.NE]
  17. “MaskGAN: Towards Diverse and Interactive Facial Image Manipulation” In IEEE Conference on Computer Vision and Pattern Recognition, 2020
  18. “Deep Learning Face Attributes in the Wild” In Proceedings of International Conference on Computer Vision (ICCV), 2015
  19. “MagFace: A universal representation for face recognition and quality assessment”, 2021
  20. “Conditional Generative Adversarial Nets”, 2014 arXiv:1411.1784 [cs.LG]
  21. “Exploiting Deep Generative Prior for Versatile Image Restoration and Manipulation”, 2020 arXiv:2003.13659 [eess.IV]
  22. Stanislav Pidhorskyi, Donald Adjeroh and Gianfranco Doretto “Adversarial Latent Autoencoders”, 2020 arXiv:2004.04467 [cs.LG]
  23. “Transforming Facial Weight of Real Images by Editing Latent Space of StyleGAN”, 2020 arXiv:2011.02606 [cs.CV]
  24. Alec Radford, Luke Metz and Soumith Chintala “Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks”, 2016 arXiv:1511.06434 [cs.LG]
  25. “Encoding in Style: a StyleGAN Encoder for Image-to-Image Translation” In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021
  26. “Image information and visual quality” In IEEE Transactions on Image Processing 15.2, 2006, pp. 430–444 DOI: 10.1109/TIP.2005.859378
  27. “Deep Learning and the Information Bottleneck Principle”, 2015 arXiv:1503.02406 [cs.LG]
  28. “Designing an Encoder for StyleGAN Image Manipulation” In arXiv preprint arXiv:2102.02766, 2021
  29. “High-Fidelity GAN Inversion for Image Attribute Editing” In arxiv:2109.06590, 2021
  30. “Image quality assessment: from error visibility to structural similarity” In IEEE Transactions on Image Processing 13.4, 2004, pp. 600–612 DOI: 10.1109/TIP.2003.819861
  31. “GAN Inversion: A Survey”, 2021 arXiv:2101.05278 [cs.CV]
  32. J. Zhou, D.L. Civco and J.A. Silander “A wavelet transform method to merge Landsat TM and SPOT panchromatic data” In International Journal of Remote Sensing 19.4 TAYLOR & FRANCIS LTD, 1998, pp. 743–757 DOI: 10.1080/014311698215973
  33. “In-Domain GAN Inversion for Real Image Editing”, 2020 arXiv:2004.00049 [cs.CV]
  34. “Generative Visual Manipulation on the Natural Image Manifold”, 2018 arXiv:1609.03552 [cs.CV]
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Wassim Kabbani (6 papers)
  2. Marcel Grimmer (10 papers)
  3. Christoph Busch (106 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.