EGAIN: Extended GAn INversion (2312.15116v1)
Abstract: Generative Adversarial Networks (GANs) have witnessed significant advances in recent years, generating increasingly higher quality images, which are non-distinguishable from real ones. Recent GANs have proven to encode features in a disentangled latent space, enabling precise control over various semantic attributes of the generated facial images such as pose, illumination, or gender. GAN inversion, which is projecting images into the latent space of a GAN, opens the door for the manipulation of facial semantics of real face images. This is useful for numerous applications such as evaluating the performance of face recognition systems. In this work, EGAIN, an architecture for constructing GAN inversion models, is presented. This architecture explicitly addresses some of the shortcomings in previous GAN inversion models. A specific model with the same name, egain, based on this architecture is also proposed, demonstrating superior reconstruction quality over state-of-the-art models, and illustrating the validity of the EGAIN architecture.
- Rameen Abdal, Yipeng Qin and Peter Wonka “Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space?”, 2019 arXiv:1904.03189 [cs.CV]
- Yuval Alaluf, Or Patashnik and Daniel Cohen-Or “ReStyle: A Residual-Based StyleGAN Encoder via Iterative Refinement” In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2021
- Lacey Best-Rowden and Anil K. Jain “Learning Face Image Quality From Human Assessments” In IEEE Transactions on Information Forensics and Security 13.12, 2018, pp. 3064–3077 DOI: 10.1109/TIFS.2018.2799585
- “Ensembling with Deep Generative Views”, 2021 arXiv:2104.14551 [cs.CV]
- “SSD-GAN: Measuring the Realness in the Spatial and Spectral Domains”, 2020 arXiv:2012.05535 [cs.CV]
- “ArcFace: Additive Angular Margin Loss for Deep Face Recognition”, 2019 arXiv:1801.07698 [cs.CV]
- Ricard Durall, Margret Keuper and Janis Keuper “Watch your Up-Convolution: CNN Based Generative Deep Neural Networks are Failing to Reproduce Spectral Distributions”, 2020 arXiv:2003.01826 [cs.CV]
- Tarik Dzanic, Karan Shah and Freddie Witherden “Fourier Spectrum Discrepancies in Deep Network Generated Images”, 2020 arXiv:1911.06465 [eess.IV]
- “SWAGAN: A Style-based Wavelet-driven Generative Model”, 2021 arXiv:2102.06108 [cs.CV]
- “Generative Adversarial Networks”, 2014 arXiv:1406.2661 [stat.ML]
- Marcel Grimmer, Raghavendra Ramachandra and Christoph Busch “Deep Face Age Progression: A Survey” In IEEE Access 9, 2021, pp. 83376–83393 DOI: 10.1109/ACCESS.2021.3085835
- “Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization”, 2017 arXiv:1703.06868 [cs.CV]
- Tero Karras, Samuli Laine and Timo Aila “A Style-Based Generator Architecture for Generative Adversarial Networks”, 2019 arXiv:1812.04948 [cs.NE]
- “Alias-Free Generative Adversarial Networks”, 2021 arXiv:2106.12423 [cs.CV]
- “Analyzing and Improving the Image Quality of StyleGAN”, 2020 arXiv:1912.04958 [cs.CV]
- “Progressive Growing of GANs for Improved Quality, Stability, and Variation”, 2018 arXiv:1710.10196 [cs.NE]
- “MaskGAN: Towards Diverse and Interactive Facial Image Manipulation” In IEEE Conference on Computer Vision and Pattern Recognition, 2020
- “Deep Learning Face Attributes in the Wild” In Proceedings of International Conference on Computer Vision (ICCV), 2015
- “MagFace: A universal representation for face recognition and quality assessment”, 2021
- “Conditional Generative Adversarial Nets”, 2014 arXiv:1411.1784 [cs.LG]
- “Exploiting Deep Generative Prior for Versatile Image Restoration and Manipulation”, 2020 arXiv:2003.13659 [eess.IV]
- Stanislav Pidhorskyi, Donald Adjeroh and Gianfranco Doretto “Adversarial Latent Autoencoders”, 2020 arXiv:2004.04467 [cs.LG]
- “Transforming Facial Weight of Real Images by Editing Latent Space of StyleGAN”, 2020 arXiv:2011.02606 [cs.CV]
- Alec Radford, Luke Metz and Soumith Chintala “Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks”, 2016 arXiv:1511.06434 [cs.LG]
- “Encoding in Style: a StyleGAN Encoder for Image-to-Image Translation” In IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021
- “Image information and visual quality” In IEEE Transactions on Image Processing 15.2, 2006, pp. 430–444 DOI: 10.1109/TIP.2005.859378
- “Deep Learning and the Information Bottleneck Principle”, 2015 arXiv:1503.02406 [cs.LG]
- “Designing an Encoder for StyleGAN Image Manipulation” In arXiv preprint arXiv:2102.02766, 2021
- “High-Fidelity GAN Inversion for Image Attribute Editing” In arxiv:2109.06590, 2021
- “Image quality assessment: from error visibility to structural similarity” In IEEE Transactions on Image Processing 13.4, 2004, pp. 600–612 DOI: 10.1109/TIP.2003.819861
- “GAN Inversion: A Survey”, 2021 arXiv:2101.05278 [cs.CV]
- J. Zhou, D.L. Civco and J.A. Silander “A wavelet transform method to merge Landsat TM and SPOT panchromatic data” In International Journal of Remote Sensing 19.4 TAYLOR & FRANCIS LTD, 1998, pp. 743–757 DOI: 10.1080/014311698215973
- “In-Domain GAN Inversion for Real Image Editing”, 2020 arXiv:2004.00049 [cs.CV]
- “Generative Visual Manipulation on the Natural Image Manifold”, 2018 arXiv:1609.03552 [cs.CV]
- Wassim Kabbani (6 papers)
- Marcel Grimmer (10 papers)
- Christoph Busch (106 papers)