Neural Knitworks: Patched Neural Implicit Representation Networks (2109.14406v2)
Abstract: Coordinate-based Multilayer Perceptron (MLP) networks, despite being capable of learning neural implicit representations, are not performant for internal image synthesis applications. Convolutional Neural Networks (CNNs) are typically used instead for a variety of internal generative tasks, at the cost of a larger model. We propose Neural Knitwork, an architecture for neural implicit representation learning of natural images that achieves image synthesis by optimizing the distribution of image patches in an adversarial manner and by enforcing consistency between the patch predictions. To the best of our knowledge, this is the first implementation of a coordinate-based MLP tailored for synthesis tasks such as image inpainting, super-resolution, and denoising. We demonstrate the utility of the proposed technique by training on these three tasks. The results show that modeling natural images using patches, rather than pixels, produces results of higher fidelity. The resulting model requires 80% fewer parameters than alternative CNN-based solutions while achieving comparable performance and training time.
- M. Tancik, P. P. Srinivasan, B. Mildenhall, S. Fridovich-Keil, N. Raghavan, U. Singhal, R. Ramamoorthi, J. T. Barron, and R. Ng, “Fourier features let networks learn high frequency functions in low dimensional domains,” NeurIPS, 2020.
- V. Sitzmann, J. N. Martel, A. W. Bergman, D. B. Lindell, and G. Wetzstein, “Implicit neural representations with periodic activation functions,” in Proc. NeurIPS, 2020.
- B. Mildenhall, P. P. Srinivasan, M. Tancik, J. T. Barron, R. Ramamoorthi, and R. Ng, “NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis,” in ECCV 2020, vol. 12346 LNCS, pp. 405–421, 2020.
- K. Zhang, G. Riegler, N. Snavely, and V. Koltun, “NeRF++: Analyzing and Improving Neural Radiance Fields,” pp. 1–9, 2020.
- A. Yu, V. Ye, M. Tancik, and A. Kanazawa, “pixelNeRF: Neural radiance fields from one or few images,” arXiv, 2020.
- M. Niemeyer and A. Geiger, “GIRAFFE: Representing scenes as compositional generative neural feature fields,” arXiv, pp. 1–12, 2020.
- A. Shocher, S. Bagon, P. Isola, and M. Irani, “InGAN: Capturing and retargeting the ’DNA’ of a natural image,” Proceedings of the IEEE International Conference on Computer Vision, vol. 2019-Octob, no. i, pp. 4491–4500, 2019.
- T. R. Shaham, T. Dekel, and T. Michaeli, “SinGAN: Learning a generative model from a single natural image,” Proceedings of the IEEE International Conference on Computer Vision, vol. 2019-Octob, pp. 4569–4579, 2019.
- T. Park, J. Y. Zhu, O. Wang, J. Lu, E. Shechtman, A. A. Efros, and R. Zhang, “Swapping Autoencoder for Deep Image Manipulation,” arXiv, no. NeurIPS, 2020.
- E. Dupont, A. Goliński, M. Alizadeh, Y. W. Teh, and A. Doucet, “COIN: COmpression with Implicit Neural representations,” pp. 1–12, 2021.
- I. Skorokhodov, S. Ignatyev, and M. Elhoseiny, “Adversarial generation of continuous images,” arXiv, 2020.
- I. Anokhin, K. Demochkin, T. Khakhulin, G. Sterkin, V. Lempitsky, and D. Korzhenkov, “Image generators with conditionally-independent pixel synthesis,” arXiv, 2020.
- J. J. Park, P. Florence, J. Straub, R. Newcombe, and S. Lovegrove, “Deepsdf: Learning continuous signed distance functions for shape representation,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2019-June, pp. 165–174, 2019.
- K. Genova, F. Cole, D. Vlasic, A. Sarna, W. T. Freeman, and T. A. Funkhouser, “Learning shape templates with structured implicit functions,” CoRR, vol. abs/1904.06447, 2019.
- M. Atzmon and Y. Lipman, “SAL: sign agnostic learning of shapes from raw data,” CoRR, vol. abs/1911.10414, 2019.
- N. Rahaman, A. Baratin, D. Arpit, F. Draxler, M. Lin, F. Hamprecht, Y. Bengio, and A. Courville, “On the spectral bias of neural networks,” in Proceedings of the 36th International Conference on Machine Learning (K. Chaudhuri and R. Salakhutdinov, eds.), vol. 97 of Proceedings of Machine Learning Research, pp. 5301–5310, PMLR, 09–15 Jun 2019.
- Y. Chen, S. Liu, and X. Wang, “Learning continuous image representation with local implicit image function,” arXiv, 2020.
- A. A. Efros and W. T. Freeman, “Image quilting for texture synthesis and transfer,” Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 2001, pp. 341–346, 2001.
- V. Kwatra, A. Schödl, I. Essa, G. Turk, and A. Bobick, “Graphcut textures: Image and video synthesis using graph cuts,” ACM Transactions on Graphics, vol. 22, no. 3, pp. 277–286, 2003.
- D. Simakov, Y. Caspi, E. Shechtman, and M. Irani, “Summarizing visual data using bidirectional similarity,” 26th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, no. ii, 2008.
- D. Glasner, S. Bagon, and M. Irani, “Super-resolution from a single image,” Proceedings of the IEEE International Conference on Computer Vision, pp. 349–356, 2009.
- M. Zontak and M. Irani, “Internal statistics of a single natural image,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 977–984, 2011.
- M. Zontak, I. Mosseri, and M. Irani, “Separating signal from noise using patch recurrence across scales,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 1195–1202, 2013.
- L. A. Gatys, A. S. Ecker, and M. Bethge, “Texture synthesis using convolutional neural networks,” Advances in Neural Information Processing Systems, vol. 2015-January, pp. 262–270, 2015.
- T. Michaeli and M. Irani, “Blind deblurring using internal patch recurrence,” Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 8691 LNCS, no. PART 3, pp. 783–798, 2014.
- A. Shocher, N. Cohen, and M. Irani, “Zero-Shot Super-Resolution Using Deep Internal Learning,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3118–3126, 2018.
- Y. Gandelsman, A. Shocher, and M. Irani, “’Double-dip’: Unsupervised image decomposition via coupled deep-image-priors,” Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 2019-June, pp. 11018–11027, 2019.
- R. Mechrez, E. Shechtman, and L. Zelnik-Manor, “Saliency driven image manipulation,” Machine Vision and Applications, vol. 30, no. 2, pp. 189–202, 2019.
- S. Bell-Kligler, A. Shocher, and M. Irani, “Blind super-resolution kernel estimation using an internal-GAN,” arXiv, no. 788535, pp. 1–10, 2019.
- I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” Advances in Neural Information Processing Systems, vol. 3, no. January, pp. 2672–2680, 2014.
- D. Ulyanov, A. Vedaldi, and V. Lempitsky, “Deep Image Prior,” International Journal of Computer Vision, vol. 128, no. 7, pp. 1867–1888, 2020.
- T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, and X. Chen, “Improved techniques for training GANs,” Advances in Neural Information Processing Systems, pp. 2234–2242, 2016.
- M. Bevilacqua, A. Roumy, C. Guillemot, and M. line Alberi Morel, “Low-complexity single-image super-resolution based on nonnegative neighbor embedding,” in Proceedings of the British Machine Vision Conference, pp. 135.1–135.10, BMVA Press, 2012.
- R. Zeyde, M. Elad, and M. Protter, “On single image scale-up using sparse-representations,” in Proceedings of the 7th International Conference on Curves and Surfaces, (Berlin, Heidelberg), p. 711–730, Springer-Verlag, 2010.