Iterative Occlusion-Aware Light Field Depth Estimation using 4D Geometrical Cues (2403.02043v1)
Abstract: Light field cameras and multi-camera arrays have emerged as promising solutions for accurately estimating depth by passively capturing light information. This is possible because the 3D information of a scene is embedded in the 4D light field geometry. Commonly, depth estimation methods extract this information relying on gradient information, heuristic-based optimisation models, or learning-based approaches. This paper focuses mainly on explicitly understanding and exploiting 4D geometrical cues for light field depth estimation. Thus, a novel method is proposed, based on a non-learning-based optimisation approach for depth estimation that explicitly considers surface normal accuracy and occlusion regions by utilising a fully explainable 4D geometric model of the light field. The 4D model performs depth/disparity estimation by determining the orientations and analysing the intersections of key 2D planes in 4D space, which are the images of 3D-space points in the 4D light field. Experimental results show that the proposed method outperforms both learning-based and non-learning-based state-of-the-art methods in terms of surface normal angle accuracy, achieving a Median Angle Error on planar surfaces, on average, 26.3\% lower than the state-of-the-art, and still being competitive with state-of-the-art methods in terms of Mean Squared Error $\vc{\times}$ 100 and Badpix 0.07.
- P. Cipresso, I. A. C. Giglioli, M. A. Raya, and G. Riva, “The past, present, and future of virtual and augmented reality research: A network and cluster analysis of the literature,” Frontiers in Psychology, vol. 9, 2018.
- S. Dargan, S. Bansal, M. Kumar, A. Mittal, and K. Kumar, “Augmented reality: A comprehensive review,” Archives of Computational Methods in Engineering, vol. 30, no. 2, pp. 1057–1080, Mar 2023.
- Raytrix-GmbH, “3D optical inspection,” https://raytrix.de/inspection/, 2018, accessed: 2018-11-30.
- D. Liu, R. Nicolescu, and R. Klette, “Bokeh effects based on stereo vision,” in Computer Analysis of Images and Patterns, G. Azzopardi and N. Petkov, Eds. Cham: Springer International Publishing, 2015, pp. 198–210.
- P. M. M. Pereira, L. A. Thomaz, L. M. N. Tavora, P. A. A. Assunção, R. Fonseca-Pinto, R. P. Paiva, and S. M. M. Faria, “Multiple instance learning using 3D features for melanoma detection,” IEEE Access, vol. 10, pp. 76 296–76 309, 2022.
- D. Scharstein and R. Szeliski, “High-accuracy stereo depth maps using structured light,” in Conference on Computer Vision and Pattern Recognition, vol. 1, 2003.
- Z. Wang and M. Menenti, “Challenges and opportunities in lidar remote sensing,” Frontiers in Remote Sensing, vol. 2, 2021.
- C. Shin, H.-G. Jeon, Y. Yoon et al., “Epinet: A fully-convolutional neural network using epipolar geometry for depth from light field images,” in Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, June 2018.
- Y.-J. Tsai, Y.-L. Liu, M. Ouhyoung, and Y.-Y. Chuang, “Attention-based view selection networks for light-field disparity estimation,” AAAI Conference on Artificial Intelligence, vol. 34, pp. 12 095–12 103, Apr. 2020.
- K. Li, J. Zhang, R. Sun, X. Zhang, and J. Gao, “Epi-based oriented relation networks for light field,” in British Machine Vision Conference, September 2020.
- W. Yan, X. Zhang, H. Chen, C. Ling, and D. Wang, “Light field depth estimation based on channel attention and edge guidance,” in 2022 China Automation Congress, 2022, pp. 2595–2600.
- L. Han, S. Zheng, Z. Shi, and M. Xia, “Exploiting sequence analysis for accurate light-field depth estimation,” IEEE Access, vol. 11, pp. 74 657–74 670, 2023.
- S. Zhang, N. Meng, and E. Y. Lam, “Unsupervised light field depth estimation via multi-view feature matching with occlusion prediction,” IEEE Transactions on Circuits and Systems for Video Technology, pp. 1–1, 2023.
- O. Johannsen, K. Honauer, B. Goldluecke et al., “A taxonomy and evaluation of dense light field depth estimation algorithms,” in Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, USA, July 2017, pp. 1795–1812.
- D. Dansereau and L. Bruton, “Gradient-based depth estimation from 4D light fields,” in International Symposium on Circuits and Systems, vol. 3, 2004, pp. III–549.
- S. Wanner and B. Goldluecke, “Globally consistent depth labeling of 4D light fields,” in Conference on Computer Vision and Pattern Recognition, Providence, USA, June 2012, pp. 41–48.
- J. Li and Z. N. Li, “Continuous depth map reconstruction from light fields,” in IEEE International Conference on Multimedia and Expo, July 2013, pp. 1–6.
- J. Li, M. Lu, and Z. Li, “Continuous depth map reconstruction from light fields,” IEEE Transactions on Image Processing, vol. 24, no. 11, pp. 3257–3265, November 2015.
- R. Lourenco, P. A. A. Assunção, L. M. N. Tavora, R. Fonseca-Pinto, and S. M. M. Faria, “Silhouette enhancement in light field disparity estimation using the structure tensor,” in International Conference on Image Processing, Athens, Greece, October 2018, pp. 2580–2584.
- R. M. Lourenco, L. M. N. Tavora, P. A. A. Assunção, L. A. Thomaz, R. Fonseca-Pinto, and S. M. M. Faria, “Enhancement of light field disparity maps by reducing the silhouette effect and plane noise,” Multidimensional Systems and Signal Processing, Jan 2022.
- N. Khan, M. H. Kim, and J. Tompkin, “Edge-aware bidirectional diffusion for dense depth estimation from light fields,” in British Machine Vision Conference, 2021.
- J. Y. Lee and R.-H. Park, “Complex-valued disparity: Unified depth model of depth from stereo, depth from focus, and depth from defocus based on the light field gradient,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 3, pp. 830–841, 2021.
- S. Zhang, H. Sheng, C. Li et al., “Robust depth estimation for light field via spinning parallelogram operator,” Computer Vision and Image Understanding, vol. 145, pp. 148–159, 2016.
- H. Schilling, M. Diebold, C. Rother, and B. Jähne, “Trust your model: Light field depth estimation with inline occlusion handling,” in Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, June 2018, pp. 4530–4538.
- M. W. Tao, S. Hadap, J. Malik, and R. Ramamoorthi, “Depth from combining defocus and correspondence using light-field cameras,” IEEE International Conference on Computer Vision, pp. 673–680, March 2013.
- H. G. Jeon, J. Park, G. Choe et al., “Accurate depth map estimation from a lenslet light field camera,” in Conference on Computer Vision and Pattern Recognition, Boston, USA, June 2015, pp. 1547–1555.
- H. Lin, C. Chen, S. Bing Kang, and J. Yu, “Depth recovery from light field using focal stack symmetry,” International Conference on Computer Vision, December 2015.
- T.-C. Wang, A. A. Efros, and R. Ramamoorthi, “Depth estimation with occlusion modeling using light-field cameras,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 38, no. 11, pp. 2170–2181, 2016.
- W. Williem and I. K. Park, “Robust light field depth estimation for noisy scene with occlusion,” in Conference on Computer Vision and Pattern Recognition, Las Vegas, USA, 2016, pp. 4396–4404.
- M. Strecke, A. Alperovich, and B. Goldluecke, “Accurate depth and normal maps from occlusion-aware focal stack symmetry,” in Computer Vision and Pattern Recognition, Honolulu, USA, July 2017, pp. 2529–2537.
- W. Williem, I. K. Park, and K. M. Lee, “Robust light field depth estimation using occlusion-noise aware data costs,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 1–1, August 2018.
- S. Ma, Z. Guo, J. Wu et al., “Occlusion-aware light field depth estimation using side window angular coherence,” Appl. Opt., vol. 60, no. 2, pp. 392–404, Jan 2021.
- K. Han, W. Xiang, E. Wang, and T. Huang, “A novel occlusion-aware vote cost for light field depth estimation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 1–1, 2021.
- M. Levoy and P. Hanrahan, “Light field rendering,” in Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, ser. SIGGRAPH ’96. New York, NY, USA: Association for Computing Machinery, 1996, p. 31–42.
- M. B. de Carvalho, C. L. Pagliari, G. O. e Alves, C. Schretter, P. Schelkens, F. Pereira, and E. A. B. da Silva, “Supporting wider baseline light fields in jpeg pleno with a novel slanted 4D-DCT coding mode,” IEEE Access, vol. 11, pp. 28 294–28 317, 2023.
- J. Bigun, “Optimal orientation detection of linear symmetry,” in IEEE First International Conf. on Computer Vision, London, Great Britain, June 1987, pp. 433–438.
- L. I. Rudin, S. Osher, and E. Fatemi, “Nonlinear total variation based noise removal algorithms,” Physica D: Nonlinear Phenomena, vol. 60, no. 1, pp. 259–268, 1992.
- M. Ruzon and C. Tomasi, “Color edge detection with the compass operator,” in Conference on Computer Vision and Pattern Recognition, vol. 2, June 1999, pp. 160–166 Vol. 2.
- W. Zhou, L. Lin, Y. Hong, Q. Li, X. Shen, and E. E. Kuruoglu, “Beyond photometric consistency: Geometry-based occlusion-aware unsupervised light field disparity estimation,” IEEE Transactions on Neural Networks and Learning Systems, pp. 1–15, July 2023.
- Y. Nakagawa, H. Uchiyama, H. Nagahara, and R.-I. Taniguchi, “Estimating surface normals with depth image gradients for fast and accurate registration,” in 2015 International Conference on 3D Vision, 2015, pp. 640–647.
- M. Gutsche, H. Schilling, M. Diebold, and C. Garbe, “Surface normal reconstruction from specular information in light field data,” in Conference on Computer Vision and Pattern Recognition Workshops, July 2017, pp. 1735–1742.
- Z. Cui, H. Sheng, D. Yang, S. Wang, R. Chen, and W. Ke, “Light field depth estimation for non-lambertian objects via adaptive cross operator,” IEEE Transactions on Circuits and Systems for Video Technology, pp. 1–1, 2023.
- D. Bertsimas and J. Tsitsiklis, “Simulated annealing,” Statistical science, vol. 8, no. 1, pp. 10–15, 1993.
- S. Kirkpatrick, C. D. Gelatt, and M. P. Vecchi, “Optimization by simulated annealing,” Science, vol. 220, no. 4598, pp. 671–680, 1983.
- M. Strecke and B. Goldluecke, “Sublabel-accurate convex relaxation with total generalized variation regularization,” in German Conference on Pattern Recognition (Proc. GCPR), 2018.