Papers
Topics
Authors
Recent
Search
2000 character limit reached

Learning Position-Aware Implicit Neural Network for Real-World Face Inpainting

Published 19 Jan 2024 in cs.CV | (2401.10537v1)

Abstract: Face inpainting requires the model to have a precise global understanding of the facial position structure. Benefiting from the powerful capabilities of deep learning backbones, recent works in face inpainting have achieved decent performance in ideal setting (square shape with $512px$). However, existing methods often produce a visually unpleasant result, especially in the position-sensitive details (e.g., eyes and nose), when directly applied to arbitrary-shaped images in real-world scenarios. The visually unpleasant position-sensitive details indicate the shortcomings of existing methods in terms of position information processing capability. In this paper, we propose an \textbf{I}mplicit \textbf{N}eural \textbf{I}npainting \textbf{N}etwork (IN$2$) to handle arbitrary-shape face images in real-world scenarios by explicit modeling for position information. Specifically, a downsample processing encoder is proposed to reduce information loss while obtaining the global semantic feature. A neighbor hybrid attention block is proposed with a hybrid attention mechanism to improve the facial understanding ability of the model without restricting the shape of the input. Finally, an implicit neural pyramid decoder is introduced to explicitly model position information and bridge the gap between low-resolution features and high-resolution output. Extensive experiments demonstrate the superiority of the proposed method in real-world face inpainting task.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. PatchMatch: A randomized correspondence algorithm for structural image editing. TOG, 28(3):24, 2009.
  2. Simultaneous structure and texture image inpainting. TIP, 12(8):882–889, 2003.
  3. CiaoSR: Continuous implicit attention-in-attention network for arbitrary-scale image super-resolution. In CVPR, 2023.
  4. Learning continuous image representation with local implicit image function. In CVPR, pages 8628–8638, 2021.
  5. Activating more pixels in image super-resolution transformer. In CVPR, pages 22367–22377, June 2023.
  6. Cascaded detail-preserving networks for super-resolution of document images. In ICDAR, pages 240–245, 2019.
  7. Generative adversarial nets. NeurIPS, 27, 2014.
  8. Neighborhood attention transformer. In CVPR, pages 6185–6194, 2023.
  9. Keys to Better Image Inpainting: Structure and texture go hand in hand. In WACV, pages 208–217, 2023.
  10. Perceptual losses for real-time style transfer and super-resolution. In ECCV, pages 694–711, 2016.
  11. Progressive growing of gans for improved quality, stability, and variation. arXiv preprint arXiv:1710.10196, 2017.
  12. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  13. Imagenet classification with deep convolutional neural networks. NeurIPS, 25, 2012.
  14. Local texture estimator for implicit representation function. In CVPR, pages 1929–1938, June 2022.
  15. MAT: Mask-aware transformer for large hole image inpainting. In CVPR, pages 10758–10768, 2022.
  16. UniFormer: Unifying convolution and self-attention for visual recognition. TPAMI, pages 1–18, 2023.
  17. Image inpainting for irregular holes using partial convolutions. In ECCV, pages 85–100, 2018.
  18. Swin transformer: Hierarchical vision transformer using shifted windows. In ICCV, pages 10012–10022, 2021.
  19. CoordFill: Efficient high-resolution image inpainting parameterized coordinate querying. In AAAI, pages 1746–1754, 2023.
  20. Nerf in the Wild: Neural radiance fields for unconstrained photo collections. In CVPR, pages 7210–7219, 2021.
  21. Which training methods for gans do actually converge? In ICML, pages 3481–3490. PMLR, 2018.
  22. EdgeConnect: Structure guided image inpainting using edge prediction. In CVPR, pages 3265–3274, 2019.
  23. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556, 2014.
  24. Resolution-robust large mask inpainting with fourier convolutions. In WACV, pages 2149–2159, 2022.
  25. Attention is all you need. NeurIPS, 30, 2017.
  26. Image quality assessment: from error visibility to structural similarity. TIP, 13(4):600–612, 2004.
  27. High-resolution image synthesis and semantic manipulation with conditional gans. In CVPR, pages 8798–8807, 2018.
  28. Towards real-world blind face restoration with generative facial prior. In CVPR, pages 9168–9178, 2021.
  29. CvT: Introducing convolutions to vision transformers. In ICCV, pages 22–31, 2021.
  30. Revisiting implicit neural representations in low-level vision. In ICLR, 2023.
  31. Texture memory-augmented deep patch-based image inpainting. TIP, 30:9112–9124, 2021.
  32. Positional encoding as spatial inductive bias in gans. In CVPR, pages 13569–13578, 2021.
  33. Diverse inpainting and editing with gan inversion. In CVPR, pages 23120–23130, 2023.
  34. Generative image inpainting with contextual attention. In CVPR, pages 5505–5514, 2018.
  35. Learning pyramid-context encoder network for high-quality image inpainting. In CVPR, pages 1486–1494, 2019.
  36. Aggregated contextual transformations for high-resolution image inpainting. TVCG, pages 3266–3280, 2022.
  37. The unreasonable effectiveness of deep features as a perceptual metric. In CVPR, pages 586–595, 2018.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (3)

Collections

Sign up for free to add this paper to one or more collections.