Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Gaussian Splatting with NeRF-based Color and Opacity (2312.13729v5)

Published 21 Dec 2023 in cs.CV

Abstract: Neural Radiance Fields (NeRFs) have demonstrated the remarkable potential of neural networks to capture the intricacies of 3D objects. By encoding the shape and color information within neural network weights, NeRFs excel at producing strikingly sharp novel views of 3D objects. Recently, numerous generalizations of NeRFs utilizing generative models have emerged, expanding its versatility. In contrast, Gaussian Splatting (GS) offers a similar render quality with faster training and inference as it does not need neural networks to work. It encodes information about the 3D objects in the set of Gaussian distributions that can be rendered in 3D similarly to classical meshes. Unfortunately, GS are difficult to condition since they usually require circa hundred thousand Gaussian components. To mitigate the caveats of both models, we propose a hybrid model Viewing Direction Gaussian Splatting (VDGS) that uses GS representation of the 3D object's shape and NeRF-based encoding of color and opacity. Our model uses Gaussian distributions with trainable positions (i.e. means of Gaussian), shape (i.e. covariance of Gaussian), color and opacity, and a neural network that takes Gaussian parameters and viewing direction to produce changes in the said color and opacity. As a result, our model better describes shadows, light reflections, and the transparency of 3D objects without adding additional texture and light components.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. Synthesizing 3d shapes via modeling multi-view depth maps and silhouettes with deep generative networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  1511–1519, 2017.
  2. Neural rgb-d surface reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  6290–6301, 2022.
  3. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp.  5855–5864, 2021.
  4. Mip-nerf 360: Unbounded anti-aliased neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  5470–5479, 2022.
  5. Blinn, J. F. A generalization of algebraic surface drawing. ACM transactions on graphics (TOG), 1(3):235–256, 1982.
  6. Efficient geometry-aware 3d generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  16123–16133, 2022.
  7. Tensorf: Tensorial radiance fields. In Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XXXII, pp.  333–350. Springer, 2022.
  8. Single-stage diffusion nerf: A unified approach to 3d generation and reconstruction. In ICCV, 2023.
  9. 3d-r2n2: A unified approach for single and multi-view 3d object reconstruction. In European conference on computer vision, pp.  628–644. Springer, 2016.
  10. Depth-supervised nerf: Fewer views and faster training for free. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  12882–12891, 2022.
  11. Accelerated generative models for 3d point cloud data. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp.  5497–5505, 2016.
  12. Plenoxels: Radiance fields without neural networks. In CVPR, pp.  5501–5510, 2022.
  13. Hierarchical surface prediction for 3d object reconstruction. In 2017 International Conference on 3D Vision (3DV), pp. 412–420. IEEE, 2017.
  14. Shape, light, and material decomposition from images using monte carlo rendering and denoising. Advances in Neural Information Processing Systems, 35:22856–22869, 2022.
  15. Gaussianshader: 3d gaussian splatting with shading functions for reflective surfaces. arXiv preprint arXiv:2311.17977, 2023.
  16. Ray tracing volume densities. ACM SIGGRAPH computer graphics, 18(3):165–174, 1984.
  17. 3d gaussian splatting for real-time radiance field rendering. ACM Transactions on Graphics, 42(4), 2023.
  18. Flexible techniques for differentiable rendering with 3d gaussians. arXiv preprint arXiv:2308.14737, 2023.
  19. Tanks and temples: Benchmarking large-scale scene reconstruction. ACM Transactions on Graphics (ToG), 36(4):1–13, 2017.
  20. Envidr: Implicit differentiable renderer with neural environment lighting. arXiv preprint arXiv:2303.13022, 2023.
  21. Neural sparse voxel fields. Advances in Neural Information Processing Systems, 33:15651–15663, 2020.
  22. Nero: Neural geometry and brdf reconstruction of reflective objects from multiview images. arXiv preprint arXiv:2305.17398, 2023.
  23. Man, P. Generating and real-time rendering of clouds. In Central European seminar on computer graphics, volume 1. Citeseer Castá-Papiernicka, Slovakia, 2006.
  24. Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020.
  25. Instant neural graphics primitives with a multiresolution hash encoding. ACM Transactions on Graphics (ToG), 41(4):1–15, 2022.
  26. Extracting triangular 3d models, materials, and lighting from images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  8280–8290, 2022.
  27. Regnerf: Regularizing neural radiance fields for view synthesis from sparse inputs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  5480–5490, 2022.
  28. Splatting with shadows. In Volume Graphics 2001: Proceedings of the Joint IEEE TCVG and Eurographics Workshop in Stony Brook, New York, USA, June 21–22, 2001, pp. 35–49. Springer, 2001.
  29. Dense depth priors for neural radiance fields from sparse input views. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  12892–12901, 2022.
  30. Deep learning 3d shape surfaces using geometry images. In European Conference on Computer Vision, pp.  223–240. Springer, 2016.
  31. Hypernetwork approach to generating point clouds. Proceedings of Machine Learning Research, 119, 2020.
  32. Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction. In CVPR, pp.  5459–5469, 2022.
  33. Block-nerf: Scalable large scene neural view synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  8248–8258, 2022.
  34. Ref-nerf: Structured view-dependent appearance for neural radiance fields. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.  5481–5490. IEEE, 2022.
  35. 4d gaussian splatting for real-time dynamic scene rendering. arXiv preprint arXiv:2310.08528, 2023.
  36. Continuous conditional random field convolution for point cloud segmentation. Pattern Recognition, 122:108357, 2022.
  37. Multiplanenerf: Neural radiance field with non-trainable representation. arXiv preprint arXiv:2305.10579, 2023.
Citations (10)

Summary

We haven't generated a summary for this paper yet.