Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Compact 3D Gaussian Representation for Radiance Field (2311.13681v2)

Published 22 Nov 2023 in cs.CV and cs.GR

Abstract: Neural Radiance Fields (NeRFs) have demonstrated remarkable potential in capturing complex 3D scenes with high fidelity. However, one persistent challenge that hinders the widespread adoption of NeRFs is the computational bottleneck due to the volumetric rendering. On the other hand, 3D Gaussian splatting (3DGS) has recently emerged as an alternative representation that leverages a 3D Gaussisan-based representation and adopts the rasterization pipeline to render the images rather than volumetric rendering, achieving very fast rendering speed and promising image quality. However, a significant drawback arises as 3DGS entails a substantial number of 3D Gaussians to maintain the high fidelity of the rendered images, which requires a large amount of memory and storage. To address this critical issue, we place a specific emphasis on two key objectives: reducing the number of Gaussian points without sacrificing performance and compressing the Gaussian attributes, such as view-dependent color and covariance. To this end, we propose a learnable mask strategy that significantly reduces the number of Gaussians while preserving high performance. In addition, we propose a compact but effective representation of view-dependent color by employing a grid-based neural field rather than relying on spherical harmonics. Finally, we learn codebooks to compactly represent the geometric attributes of Gaussian by vector quantization. With model compression techniques such as quantization and entropy coding, we consistently show over 25$\times$ reduced storage and enhanced rendering speed, while maintaining the quality of the scene representation, compared to 3DGS. Our work provides a comprehensive framework for 3D scene representation, achieving high performance, fast training, compactness, and real-time rendering. Our project page is available at https://maincold2.github.io/c3dgs/.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (48)
  1. Neural point-based graphics. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXII 16, pages 696–712. Springer, 2020.
  2. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5855–5864, 2021.
  3. Mip-nerf 360: Unbounded anti-aliased neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5470–5479, 2022.
  4. Zip-nerf: Anti-aliased grid-based neural radiance fields. arXiv preprint arXiv:2304.06706, 2023.
  5. Estimating or propagating gradients through stochastic neurons for conditional computation. arXiv preprint arXiv:1308.3432, 2013.
  6. Efficient geometry-aware 3d generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16123–16133, 2022.
  7. Tensorf: Tensorial radiance fields. In European Conference on Computer Vision, 2022.
  8. Depth-supervised nerf: Fewer views and faster training for free. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 12882–12891, 2022.
  9. Fast dynamic radiance fields with time-aware neural voxels. In SIGGRAPH Asia 2022 Conference Papers, 2022.
  10. Plenoxels: Radiance fields without neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5501–5510, 2022.
  11. K-planes: Explicit radiance fields in space, time, and appearance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12479–12488, 2023.
  12. Dynamic view synthesis from dynamic monocular video. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 5712–5721, 2021.
  13. Strivec: Sparse tri-vector radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 17569–17579, 2023.
  14. Robert Gray. Vector quantization. IEEE Assp Magazine, 1(2):4–29, 1984.
  15. Multiscale tensor decomposition and rendering equation encoding for view synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4232–4241, 2023.
  16. Deep blending for free-viewpoint image-based rendering. ACM Transactions on Graphics (ToG), 37(6):1–15, 2018.
  17. 3d gaussian splatting for real-time radiance field rendering. ACM Transactions on Graphics (ToG), 42(4):1–14, 2023.
  18. Tanks and temples: Benchmarking large-scale scene reconstruction. ACM Transactions on Graphics (ToG), 36(4):1–13, 2017.
  19. Point-based neural rendering with per-view optimization. In Computer Graphics Forum, pages 29–43, 2021.
  20. Neural point catacaustics for novel-view synthesis of reflections. ACM Transactions on Graphics (TOG), 41(6):1–15, 2022.
  21. Pulsar: Efficient sphere-based neural rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 1440–1449, 2021.
  22. Neural 3d video synthesis from multi-view video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5521–5531, 2022.
  23. Neural scene flow fields for space-time view synthesis of dynamic scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 6498–6508, 2021.
  24. Devrf: Fast deformable voxel radiance fields for dynamic scenes. Advances in Neural Information Processing Systems, 35:36762–36775, 2022.
  25. Neural sparse voxel fields. Advances in Neural Information Processing Systems, pages 15651–15663, 2020.
  26. Neural rerendering in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
  27. Nerf: Representing scenes as neural radiance fields for view synthesis. In European Conference on Computer Vision, page 405–421, 2020.
  28. Nerf in the dark: High dynamic range view synthesis from noisy raw images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 16190–16199, 2022.
  29. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph., 41(4), 2022.
  30. Mip-grid: Anti-aliased grid representations for neural radiance fields. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  31. Regnerf: Regularizing neural radiance fields for view synthesis from sparse inputs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5480–5490, 2022.
  32. D-nerf: Neural radiance fields for dynamic scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10318–10327, 2021.
  33. Kilonerf: Speeding up neural radiance fields with thousands of tiny mlps. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 14335–14345, 2021.
  34. Masked wavelet representation for compact neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 20680–20690, 2023.
  35. Dense depth priors for neural radiance fields from sparse input views. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 12892–12901, 2022.
  36. Structure-from-motion revisited. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
  37. Binary radiance fields. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  38. Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5459–5469, 2022.
  39. Variable bitrate neural fields. In ACM SIGGRAPH 2022 Conference Proceedings, 2022.
  40. Block-nerf: Scalable large scene neural view synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 8248–8258, 2022.
  41. Ibrnet: Learning multi-view image-based rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4690–4699, 2021.
  42. Synsin: End-to-end view synthesis from a single image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
  43. Sinnerf: Training neural radiance fields on complex scenes from a single image. In European Conference on Computer Vision, pages 736–753, 2022a.
  44. Point-nerf: Point-based neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 5438–5448, 2022b.
  45. Differentiable surface splatting for point-based geometry processing. ACM Transactions on Graphics (TOG), 38(6):1–14, 2019.
  46. pixelnerf: Neural radiance fields from one or few images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4578–4587, 2021.
  47. Soundstream: An end-to-end neural audio codec. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 30:495–507, 2021.
  48. Differentiable point-based radiance fields for efficient view synthesis. In SIGGRAPH Asia 2022 Conference Papers, pages 1–12, 2022.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Joo Chan Lee (10 papers)
  2. Daniel Rho (13 papers)
  3. Xiangyu Sun (16 papers)
  4. Jong Hwan Ko (30 papers)
  5. Eunbyung Park (42 papers)
Citations (107)

Summary

  • The paper introduces a compact 3D Gaussian framework that filters redundant data using a novel masking strategy and grid-based neural fields.
  • The method achieves over tenfold storage reduction while boosting rendering speed and maintaining high reconstruction quality across varied datasets.
  • The approach leverages vector quantization for geometric attributes to efficiently balance performance and data compression in neural rendering.

Introduction to Neural Radiance Fields

Neural Radiance Fields (NeRF) have emerged as a transformative technology in the field of computer vision and graphics. With a focus on synthesizing photorealistic 3D scenes, NeRFs offer an impressive ability to construct high-fidelity images from a set of 2D inputs. Despite their prowess, they are computationally intensive primarily due to the requirement of volumetric rendering, which demands dense point sampling along rays to render a single pixel. This has presented a significant barrier to their adoption for real-time applications on devices with limited computational power.

Advancements in 3D Scene Representation

Addressing the need for faster rendering without compromising image quality, 3D Gaussian splatting (3DGS) has come to the fore as an alternative representation strategy. 3DGS replaces the volumetric rendering with a rasterization-based technique, which together with optimized computation, achieves exceptional rendering speeds. The caveat, however, is that 3DGS relies on a large number of 3D Gaussians, leading to exorbitant memory and storage demands – a key challenge that the latest research efforts seek to tackle.

Compact 3D Gaussian Framework

Recent developments in the field have introduced a compact 3D Gaussian framework designed to significantly enhance the efficiency of memory and storage while maintaining high-quality scene reconstructions. The approach embraces two main objectives: minimizing the number of required Gaussians and compressing their attributes. A novel masking strategy is employed to filter out redundant Gaussians, and a technique involving a grid-based neural field circumvents the traditional reliance on spherical harmonics for storing color data. Moreover, the utilization of vector quantization codebooks for geometric attributes rounds out the components of this compact representation. The resultant framework strikes a balance between performance and efficiency, demonstrating a remarkable reduction in storage and an increase in rendering speed in extensive experimental validations.

Results and Benchmarks

The testing of this compact representation across diverse datasets, encompassing both real and synthetic scenes, illustrates a consistent improvement in storage efficiency and rendering speed, achieving over tenfold storage reduction and enhanced rendering performance. Notably, on real-world datasets, the approach sets a new benchmark in reconstruction quality while boasting significant storage efficiency and an increase in rendering speed. All these advancements are integrated into a comprehensive framework that provides a pathway towards high-performance, fast-training, compact, and real-time rendering of 3D scenes.

Conclusion and Future Implications

The development of this compact 3D Gaussian representation addresses critical issues in memory and storage that have previously limited the practical applications of NeRFs and 3DGS. By facilitating high-quality rendering at an accelerated pace and with reduced storage needs, the framework is poised to accelerate the adoption of neural rendering in interactive applications and beyond. The project's contributions mark a significant stride towards overcoming computational barriers and enhancing the accessibility of advanced 3D scene rendering technologies.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com

HackerNews