Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

2D Gaussian Splatting for Geometrically Accurate Radiance Fields (2403.17888v2)

Published 26 Mar 2024 in cs.CV and cs.GR

Abstract: 3D Gaussian Splatting (3DGS) has recently revolutionized radiance field reconstruction, achieving high quality novel view synthesis and fast rendering speed without baking. However, 3DGS fails to accurately represent surfaces due to the multi-view inconsistent nature of 3D Gaussians. We present 2D Gaussian Splatting (2DGS), a novel approach to model and reconstruct geometrically accurate radiance fields from multi-view images. Our key idea is to collapse the 3D volume into a set of 2D oriented planar Gaussian disks. Unlike 3D Gaussians, 2D Gaussians provide view-consistent geometry while modeling surfaces intrinsically. To accurately recover thin surfaces and achieve stable optimization, we introduce a perspective-correct 2D splatting process utilizing ray-splat intersection and rasterization. Additionally, we incorporate depth distortion and normal consistency terms to further enhance the quality of the reconstructions. We demonstrate that our differentiable renderer allows for noise-free and detailed geometry reconstruction while maintaining competitive appearance quality, fast training speed, and real-time rendering.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (73)
  1. Neural point-based graphics. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXII 16. Springer, 696–712.
  2. Mip-NeRF: A Multiscale Representation for Anti-Aliasing Neural Radiance Fields. ICCV (2021).
  3. Mip-nerf 360: Unbounded anti-aliased neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 5470–5479.
  4. Mip-NeRF 360: Unbounded Anti-Aliased Neural Radiance Fields. CVPR (2022).
  5. Zip-NeRF: Anti-Aliased Grid-Based Neural Radiance Fields. ICCV (2023).
  6. High-quality surface splatting on today’s GPUs. In Proceedings Eurographics/IEEE VGTC Symposium Point-Based Graphics, 2005. IEEE, 17–141.
  7. TensoRF: Tensorial Radiance Fields. In European Conference on Computer Vision (ECCV).
  8. NeuSG: Neural Implicit Surface Reconstruction with 3D Gaussian Splatting Guidance. arXiv preprint arXiv:2312.00846 (2023).
  9. Mobilenerf: Exploiting the polygon rasterization pipeline for efficient neural field rendering on mobile architectures. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 16569–16578.
  10. NeuRBF: A Neural Fields Representation with Adaptive Radial Basis Functions. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 4182–4194.
  11. Plenoxels: Radiance Fields without Neural Networks. In CVPR.
  12. Geo-Neus: Geometry-Consistent Neural Implicit Surfaces Learning for Multi-view Reconstruction. Advances in Neural Information Processing Systems (NeurIPS) (2022).
  13. Relightable 3D Gaussian: Real-time Point Cloud Relighting with BRDF Decomposition and Ray Tracing. arXiv:2311.16043 (2023).
  14. Antoine Guédon and Vincent Lepetit. 2023. SuGaR: Surface-Aligned Gaussian Splatting for Efficient 3D Mesh Reconstruction and High-Quality Mesh Rendering. arXiv preprint arXiv:2311.12775 (2023).
  15. Baking neural radiance fields for real-time view synthesis. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 5875–5884.
  16. Tri-MipRF: Tri-Mip Representation for Efficient Anti-Aliasing Neural Radiance Fields. In ICCV.
  17. Eldar Insafutdinov and Alexey Dosovitskiy. 2018. Unsupervised learning of shape and pose with differentiable point clouds. Advances in neural information processing systems 31 (2018).
  18. Large scale multi-view stereopsis evaluation. In Proceedings of the IEEE conference on computer vision and pattern recognition. 406–413.
  19. GaussianShader: 3D Gaussian Splatting with Shading Functions for Reflective Surfaces. arXiv preprint arXiv:2311.17977 (2023).
  20. Michael Kazhdan and Hugues Hoppe. 2013. Screened poisson surface reconstruction. ACM Transactions on Graphics (ToG) 32, 3 (2013), 1–13.
  21. 3D Gaussian Splatting for Real-Time Radiance Field Rendering. ACM Transactions on Graphics 42, 4 (July 2023). https://repo-sam.inria.fr/fungraph/3d-gaussian-splatting/
  22. Leonid Keselman and Martial Hebert. 2022. Approximate differentiable rendering with algebraic surfaces. In European Conference on Computer Vision. Springer, 596–614.
  23. Leonid Keselman and Martial Hebert. 2023. Flexible techniques for differentiable rendering with 3d gaussians. arXiv preprint arXiv:2308.14737 (2023).
  24. Tanks and Temples: Benchmarking Large-Scale Scene Reconstruction. ACM Transactions on Graphics 36, 4 (2017).
  25. Point-Based Neural Rendering with Per-View Optimization. In Computer Graphics Forum, Vol. 40. Wiley Online Library, 29–43.
  26. Christoph Lassner and Michael Zollhofer. 2021. Pulsar: Efficient sphere-based neural rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 1440–1449.
  27. Neuralangelo: High-Fidelity Neural Surface Reconstruction. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
  28. GS-IR: 3D Gaussian Splatting for Inverse Rendering. arXiv preprint arXiv:2311.16473 (2023).
  29. Neural Sparse Voxel Fields. NeurIPS (2020).
  30. Dynamic 3D Gaussians: Tracking by Persistent Dynamic View Synthesis. In 3DV.
  31. Occupancy Networks: Learning 3D Reconstruction in Function Space. In Conference on Computer Vision and Pattern Recognition (CVPR).
  32. NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis. In ECCV.
  33. Nerf: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 65, 1 (2021), 99–106.
  34. Instant Neural Graphics Primitives with a Multiresolution Hash Encoding. ACM Trans. Graph. 41, 4, Article 102 (July 2022), 15 pages.
  35. Differentiable Volumetric Rendering: Learning Implicit 3D Representations without 3D Supervision. In Conference on Computer Vision and Pattern Recognition (CVPR).
  36. UNISURF: Unifying Neural Implicit Surfaces and Radiance Fields for Multi-View Reconstruction. In International Conference on Computer Vision (ICCV).
  37. DeepSDF: Learning Continuous Signed Distance Functions for Shape Representation. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
  38. Surfels: Surface elements as rendering primitives. In Proceedings of the 27th annual conference on Computer graphics and interactive techniques. 335–342.
  39. GaussianAvatars: Photorealistic Head Avatars with Rigged 3D Gaussians. arXiv preprint arXiv:2312.02069 (2023).
  40. KiloNeRF: Speeding up Neural Radiance Fields with Thousands of Tiny MLPs. In International Conference on Computer Vision (ICCV).
  41. Merf: Memory-efficient radiance fields for real-time view synthesis in unbounded scenes. ACM Transactions on Graphics (TOG) 42, 4 (2023), 1–12.
  42. Adop: Approximate differentiable one-pixel point rendering. ACM Transactions on Graphics (ToG) 41, 4 (2022), 1–14.
  43. Johannes Lutz Schönberger and Jan-Michael Frahm. 2016. Structure-from-Motion Revisited. In Conference on Computer Vision and Pattern Recognition (CVPR).
  44. Pixelwise View Selection for Unstructured Multi-View Stereo. In European Conference on Computer Vision (ECCV).
  45. Surfelmeshing: Online surfel-based mesh reconstruction. IEEE transactions on pattern analysis and machine intelligence 42, 10 (2019), 2494–2507.
  46. GIR: 3D Gaussian Inverse Rendering for Relightable Scene Factorization. Arxiv (2023). arXiv:2312.05133
  47. GPU-based ray-casting of quadratic surfaces.. In PBG@ SIGGRAPH. 59–65.
  48. Direct Voxel Grid Optimization: Super-fast Convergence for Radiance Fields Reconstruction. In CVPR.
  49. Improved Direct Voxel Grid Optimization for Radiance Fields Reconstruction. arxiv cs.GR 2206.05085 (2022).
  50. John Vince. 2008. Geometric algebra for computer graphics. Springer Science & Business Media.
  51. NeuS: Learning Neural Implicit Surfaces by Volume Rendering for Multi-view Reconstruction. Advances in Neural Information Processing Systems 34 (2021), 27171–27183.
  52. NeuS2: Fast Learning of Neural Implicit Surfaces for Multi-view Reconstruction. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV).
  53. A hardware architecture for surface splatting. ACM Transactions on Graphics (TOG) 26, 3 (2007), 90–es.
  54. ElasticFusion: Real-time dense SLAM and light source estimation. The International Journal of Robotics Research 35, 14 (2016), 1697–1716.
  55. SynSin: End-to-end View Synthesis from a Single Image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
  56. PhysGaussian: Physics-Integrated 3D Gaussians for Generative Dynamics. arXiv preprint arXiv:2311.12198 (2023).
  57. Street Gaussians for Modeling Dynamic Urban Scenes. (2023).
  58. MVSNet: Depth Inference for Unstructured Multi-view Stereo. European Conference on Computer Vision (ECCV) (2018).
  59. Volume rendering of neural implicit surfaces. Advances in Neural Information Processing Systems 34 (2021), 4805–4815.
  60. BakedSDF: Meshing Neural SDFs for Real-Time View Synthesis. arXiv (2023).
  61. Multiview Neural Surface Reconstruction by Disentangling Geometry and Appearance. Advances in Neural Information Processing Systems 33 (2020).
  62. Differentiable surface splatting for point-based geometry processing. ACM Transactions on Graphics (TOG) 38, 6 (2019), 1–14.
  63. PlenOctrees for Real-time Rendering of Neural Radiance Fields. In ICCV.
  64. SDFStudio: A Unified Framework for Surface Reconstruction. https://github.com/autonomousvision/sdfstudio
  65. Mip-Splatting: Alias-free 3D Gaussian Splatting. Conference on Computer Vision and Pattern Recognition (CVPR) (2024).
  66. Zehao Yu and Shenghua Gao. 2020. Fast-MVSNet: Sparse-to-Dense Multi-View Stereo With Learned Propagation and Gauss-Newton Refinement. In Conference on Computer Vision and Pattern Recognition (CVPR).
  67. MonoSDF: Exploring Monocular Geometric Cues for Neural Implicit Surface Reconstruction. Advances in Neural Information Processing Systems (NeurIPS) (2022).
  68. NeRF++: Analyzing and Improving Neural Radiance Fields. arXiv:2010.07492 (2020).
  69. Open3D: A Modern Library for 3D Data Processing. arXiv:1801.09847 (2018).
  70. Drivable 3D Gaussian Avatars. (2023). arXiv:2311.08581 [cs.CV]
  71. EWA volume splatting. In Proceedings Visualization, 2001. VIS’01. IEEE, 29–538.
  72. Surface splatting. In Proceedings of the 28th annual conference on Computer graphics and interactive techniques. 371–378.
  73. Perspective accurate splatting. In Proceedings-Graphics Interface. 247–254.
Citations (209)

Summary

  • The paper proposes a perspective-accurate 2D splatting process that recovers thin surfaces using efficient ray-splat intersection and rasterization.
  • It introduces depth distortion and normal consistency regularization, ensuring smoother and more accurate geometry reconstructions.
  • The method achieves state-of-the-art view synthesis and real-time rendering, enabling rapid training and practical deployment in complex scenes.

2D Gaussian Splatting Enhances Geometric Accuracy in Radiance Fields Reconstruction

Introduction to 2D Gaussian Splatting

The paper introduces 2D Gaussian Splatting (2DGS), an innovative approach for modeling and reconstructing geometrically accurate radiance fields from multi-view images. By collapsing the 3D volume into a set of 2D oriented planar Gaussian disks, this method addresses the limitations of 3D Gaussian Splatting (3DGS) in representing thin surfaces accurately due to the multi-view inconsistent nature of 3D Gaussians. Through perspective-accurate 2D splatting, depth distortion, and normal consistency terms, 2DGS achieves noise-free and detailed geometry reconstruction, competitive appearance quality, rapid training speeds, and real-time rendering.

Key Contributions

The significant contributions of this work are three-fold:

  • The proposal of a perspective-accurate 2D splatting process, leveraging ray-splat intersection and rasterization, to accurately recover thin surfaces and achieve stable optimization.
  • The introduction of depth distortion and normal consistency regularization terms aimed at improving the quality of the reconstructions.
  • Demonstrated state-of-the-art geometry reconstruction and novel view synthesis results when compared with existing explicit representation methods, alongside showcasing time-efficient training and real-time rendering capabilities.

Methodology

2D Gaussian Primitive Modeling: The core innovation lies in representing a 3D scene with "flat" 2D Gaussian primitives, each defining an oriented elliptical disk in 3D space. This method facilitates accurate geometry representation during rendering by employing explicit ray-splat intersection, resulting in perspective-accurate splatting.

Differentiable 2D Splatting Process: The paper outlines a differentiable renderer capable of executing a perspective-accurate 2D splatting process. This is achieved through efficient ray-splat intersection computations and a volumetric integration method for rasterization.

Regularization for Improved Reconstruction: The research introduces two regularization losses - depth distortion and normal consistency. These are instrumental in achieving smoother surfaces by compelling the 2D primitives to distribute within a tight range along the ray and aligning the geometries defined by depth and normals.

Implications and Future Developments

Theoretical Implications: This work significantly advances the understanding and capabilities of surface representations within radiance field reconstructions. By shifting from 3D to 2D Gaussian primitives, it addresses the inherent inconsistencies of 3DGS and highlights the potential of 2D representations in capturing geometrically accurate surfaces.

Practical Relevance: 2DGS opens new avenues for real-time rendering and detailed geometry reconstruction in computer graphics and vision applications, facilitating rapid training and deployment even on resource-constrained environments.

Speculation on Future Developments: The proposed method lays a foundation for further exploration into 2D-based splatting techniques. Future work might focus on enhancing the adaptability of 2DGS to more complex surface types, including semi-transparent or highly reflective materials. Moreover, the integration of 2DGS with advanced machine learning and deep learning frameworks could unlock unprecedented performance levels in radiance field reconstruction and beyond.

Conclusion

2D Gaussian Splatting presents a promising direction in the quest for geometrically accurate radiance field reconstruction. Its ability to accurately model thin surfaces, alongside the proposed perspective-accurate splatting process and regularization terms, propels it ahead of existing methodologies. As researchers continue to explore and refine this approach, 2DGS is poised to become a cornerstone in the future of realistic rendering and geometry reconstruction.

Youtube Logo Streamline Icon: https://streamlinehq.com