Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
144 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SMERF: Streamable Memory Efficient Radiance Fields for Real-Time Large-Scene Exploration (2312.07541v3)

Published 12 Dec 2023 in cs.CV and cs.GR

Abstract: Recent techniques for real-time view synthesis have rapidly advanced in fidelity and speed, and modern methods are capable of rendering near-photorealistic scenes at interactive frame rates. At the same time, a tension has arisen between explicit scene representations amenable to rasterization and neural fields built on ray marching, with state-of-the-art instances of the latter surpassing the former in quality while being prohibitively expensive for real-time applications. In this work, we introduce SMERF, a view synthesis approach that achieves state-of-the-art accuracy among real-time methods on large scenes with footprints up to 300 m$2$ at a volumetric resolution of 3.5 mm$3$. Our method is built upon two primary contributions: a hierarchical model partitioning scheme, which increases model capacity while constraining compute and memory consumption, and a distillation training strategy that simultaneously yields high fidelity and internal consistency. Our approach enables full six degrees of freedom (6DOF) navigation within a web browser and renders in real-time on commodity smartphones and laptops. Extensive experiments show that our method exceeds the current state-of-the-art in real-time novel view synthesis by 0.78 dB on standard benchmarks and 1.78 dB on large scenes, renders frames three orders of magnitude faster than state-of-the-art radiance field models, and achieves real-time performance across a wide variety of commodity devices, including smartphones. We encourage readers to explore these models interactively at our project website: https://smerf-3d.github.io.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (82)
  1. Neural point-based graphics. Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXII 16, 2020.
  2. Learning neural light fields with ray-space embedding networks. CVPR, 2022.
  3. DSSIM: a structural similarity index for floating-point data, 2023.
  4. Mip-NeRF: A multiscale representation for anti-aliasing neural radiance fields. ICCV, 2021.
  5. Mip-NeRF 360: Unbounded anti-aliased neural radiance fields. CVPR, 2022.
  6. Zip-NeRF: Anti-aliased grid-based neural radiance fields. ICCV, 2023.
  7. Unstructured Lumigraph rendering. SIGGRAPH, 2001.
  8. Real-time neural light field on mobile devices. CVPR, 2023.
  9. Efficient geometry-aware 3D generative adversarial networks. CVPR, 2022.
  10. TensoRF: Tensorial radiance fields. ECCV, 2022.
  11. MobileNeRF: Exploiting the polygon rasterization pipeline for efficient neural field rendering on mobile architectures. CVPR, 2023.
  12. Unstructured light fields. Comput. Graph. Forum, 2012.
  13. Efficient view-dependent image-based rendering with projective texture-mapping. EGSR, 1998.
  14. LightGaussian: Unbounded 3d Gaussian compression with 15x reduction and 200+ fps. arXiv, 2023.
  15. DeepView: View synthesis with learned gradient descent. CVPR, 2019.
  16. FastNeRF: High-fidelity neural rendering at 200fps. ICCV, 2021.
  17. Knowledge distillation: A survey. IJCV, 2021.
  18. LightSpeed: Light and fast neural light fields on mobile devices. arXiv, 2023a.
  19. MCNeRF: Monte carlo rendering and denoising for real-time nerfs. SIGGRAPH Asia, 2023b.
  20. Deep blending for free-viewpoint image-based rendering. SIGGRAPH Asia, 2018.
  21. Baking neural radiance fields for real-time view synthesis. ICCV, 2021.
  22. Distilling the knowledge in a neural network. arXiv:1503.02531, 2015.
  23. Tri-miprf: Tri-mip representation for efficient anti-aliasing neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 19774–19783, 2023.
  24. Alignerf: High-fidelity neural radiance fields via alignment-aware training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 46–55, 2023.
  25. 3D Gaussian splatting for real-time radiance field rendering. SIGGRAPH, 2023.
  26. Point-based neural rendering with per-view optimization. Computer Graphics Forum, 2021.
  27. AdaNeRF: Adaptive sampling for real-time rendering of neural radiance fields. ECCV, 2022.
  28. Kevin Kwok. splat, 2023. https://github.com/antimatter15/splat.
  29. Barf: Bundle-adjusting neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5741–5751, 2021.
  30. Real-time neural rasterization for large scenes. CVPR, 2023.
  31. Lookingood: Enhancing performance capture with real-time neural re-rendering. arXiv preprint arXiv:1811.05029, 2018.
  32. NeRF in the wild: Neural radiance fields for unconstrained photo collections. CVPR, 2021.
  33. Nelson Max. Optical models for direct volume rendering. IEEE TVCG, 1995.
  34. Progressively optimized local radiance fields for robust view synthesis. CVPR, 2023.
  35. Local light field fusion: Practical view synthesis with prescriptive sampling guidelines. ACM Transactions on Graphics, 2019.
  36. NeRF: Representing scenes as neural radiance fields for view synthesis. ECCV, 2020.
  37. Instant neural graphics primitives with a multiresolution hash encoding. SIGGRAPH, 2022.
  38. DONeRF: Towards real-time rendering of compact neural radiance fields using depth oracle networks. Computer Graphics Forum, 2021.
  39. RegNeRF: Regularizing neural radiance fields for view synthesis from sparse inputs. CVPR, 2022.
  40. Nerfies: Deformable neural radiance fields. ICCV, 2021.
  41. Camp: Camera preconditioning for neural radiance fields. SIGGRAPH Asia, 2023.
  42. Soft 3D reconstruction for view synthesis. SIGGRAPH Asia, 2017.
  43. Surfels: Surface elements as rendering primitives. SIGGRAPH, 2000.
  44. Floaters no more: Radiance field gradient scaling for improved near-camera training. Eurographics Symposium on Rendering, 2023.
  45. NeRFMeshing: Distilling neural radiance fields into geometrically-accurate 3d meshes. 3DV, 2023.
  46. DeRF: Decomposed radiance fields. CVPR, 2019.
  47. KiloNeRF: Speeding up neural radiance fields with thousands of tiny MLPs. ICCV, 2021.
  48. MERF: Memory-efficient radiance fields for real-time view synthesis in unbounded scenes. SIGGRAPH, 2023.
  49. Urban radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12932–12942, 2022.
  50. GANeRF: Leveraging discriminators to optimize neural radiance fields. SIGGRAPH Asia, 2023.
  51. Re-ReND: Real-time rendering of NeRFs across devices. CVPR, 2023.
  52. ADOP: Approximate differentiable one-pixel point rendering. SIGGRAPH, 2022.
  53. Structure-from-motion revisited. CVPR, 2016.
  54. Sc-nerf: Self-correcting neural radiance field with sparse views. arXiv preprint arXiv:2309.05028, 2023.
  55. NeRV: Neural reflectance and visibility fields for relighting and view synthesis. CVPR, 2021.
  56. Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction. CVPR, 2022.
  57. Surface modeling with oriented particle systems. SIGGRAPH, 1992.
  58. Block-NeRF: Scalable large scene neural view synthesis. CVPR, 2022.
  59. Advances in neural rendering. Computer Graphics Forum, 2022.
  60. Mega-NERF: Scalable construction of large-scale nerfs for virtual fly-throughs. CVPR, 2022.
  61. HybridNeRF: Efficient neural rendering via adaptive volumetric surfaces. arXiv, 2023.
  62. Attention is all you need. NeurIPS, 2017.
  63. Learning neural duplex radiance fields for real-time view synthesis. CVPR, 2023.
  64. R2L: Distilling neural radiance field to neural light field for efficient novel view synthesis. ECCV, 2022.
  65. NeuS: Learning neural implicit surfaces by volume rendering for multi-view reconstruction. NeurIPS, 2021.
  66. Adaptive shells for efficient neural radiance field rendering. SIGGRAPH Asia, 2023.
  67. Synsin: End-to-end view synthesis from a single image. CVPR, 2020.
  68. DIVeR: Real-time and accurate neural radiance fields with deterministic integration for volume rendering. CVPR, 2022a.
  69. ReconFusion: 3d reconstruction with diffusion priors. arXiv, 2023a.
  70. Scalable neural indoor scene rendering. ACM TOG, 2022b.
  71. ScaNeRF: Scalable bundle-adjusting neural radiance fields for large-scale scene rendering. ACM Transactions on Graphics (TOG), 2023b.
  72. VR-NeRF: High-fidelity virtualized walkable spaces. SIGGRAPH Asia, 2023.
  73. PlenVDB: Memory efficient vdb-based radiance fields for fast training and rendering. In CVPR, 2023.
  74. Freenerf: Improving few-shot neural rendering with free frequency regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8254–8263, 2023.
  75. BakedSDF: Meshing neural SDFs for real-time view synthesis. SIGGRAPH, 2023.
  76. PlenOctrees for real-time rendering of neural radiance fields. ICCV, 2021.
  77. Plenoxels: Radiance fields without neural networks. CVPR, 2022.
  78. NeRF++: Analyzing and improving neural radiance fields. arXiv:2010.07492, 2020.
  79. Differentiable point-based radiance fields for efficient view synthesis. SIGGRAPH Asia, 2022.
  80. Stereo magnification: Learning view synthesis using multiplane images. SIGGRAPH, 2018.
  81. Surface splatting. SIGGRAPH, 2001.
  82. Jakub Červený. gsplat — 3D gaussian splatting WebGL viewer, 2023. https://gsplat.tech/.
Citations (30)

Summary

  • The paper introduces SMERF, a novel method that splits large scenes into hierarchical submodels to enable memory-efficient, real-time rendering.
  • It employs a distillation training strategy where a high-fidelity teacher NeRF guides a student model, ensuring high-quality rendering and smooth transitions.
  • Experimental results demonstrate that SMERF achieves or surpasses current real-time methods, making photorealistic scene exploration accessible on everyday devices.

Introduction

Recent technological advancements have seen a significant leap in real-time view synthesis quality and speed. One of the fundamental challenges in this domain is reconciling highly detailed scene representations with the demands of interactive frame rates, especially for large, complex scenes. While explicit representations like meshes and point clouds have traditionally been used for this purpose, neural fields, particularly Neural Radiance Fields (NeRFs), have shown remarkable results in rendering photorealistic scenes—albeit at the cost of high computational resources, making them less feasible for real-time applications.

SMERF: A Scalable Approach to Radiance Fields

Addressing the need for real-time rendering capabilities, the newly developed method titled SMERF (Streamable Memory Efficient Radiance Fields) provides a scalable solution to view synthesis of large-scale scenes. SMERF utilizes a novel hierarchical model architecture consisting of multiple submodels that increase rendering capacity while constraining resource usage. Each submodel is specialized for a specific region of the scene, thus requiring only a fraction of the submodels to be active during rendering.

Additionally, SMERF applies a distillation training strategy whereby the model learns from a "teacher" NeRF that has already mastered rendering the scene at high fidelity, but at a slower pace not suitable for real-time applications. This allows the "student" SMERF model to adopt the same high-quality rendering capabilities and maintain coherence throughout the scene when transitioning from one submodel to another.

Real-time Rendering Across Devices

SMERF's ingenuity lies in its ability to run seamlessly on a wide variety of devices, including those with limited resources like smartphones and laptops. Experimental results indicate that SMERF not only meets but in some cases surpasses the current best real-time methods in terms of view synthesis fidelity—closing in on the quality of offline methods that are considered state-of-the-art. Remarkably, the method delivers these results while keeping rendering times in a real-time frame and adhering to memory constraints imposed by everyday consumer electronics.

Conclusion

The innovation presented by SMERF ushers in a new era of possibility for real-time exploration of large-scale 3D scenes. With its memory-efficient rendering that doesn't sacrifice image quality or rendering speed, SMERF stands as a significant step forward in the field of interactive 3D graphics and virtual exploration. Whether for gaming, virtual tours, or other interactive applications, SMERF offers a powerful tool to render detailed and immersive 3D environments in real time on standard consumer hardware.

Github Logo Streamline Icon: https://streamlinehq.com

GitHub

Youtube Logo Streamline Icon: https://streamlinehq.com