Neural Scene Chronology (2306.07970v1)
Abstract: In this work, we aim to reconstruct a time-varying 3D model, capable of rendering photo-realistic renderings with independent control of viewpoint, illumination, and time, from Internet photos of large-scale landmarks. The core challenges are twofold. First, different types of temporal changes, such as illumination and changes to the underlying scene itself (such as replacing one graffiti artwork with another) are entangled together in the imagery. Second, scene-level temporal changes are often discrete and sporadic over time, rather than continuous. To tackle these problems, we propose a new scene representation equipped with a novel temporal step function encoding method that can model discrete scene-level content changes as piece-wise constant functions over time. Specifically, we represent the scene as a space-time radiance field with a per-image illumination embedding, where temporally-varying scene changes are encoded using a set of learned step functions. To facilitate our task of chronology reconstruction from Internet imagery, we also collect a new dataset of four scenes that exhibit various changes over time. We demonstrate that our method exhibits state-of-the-art view synthesis results on this dataset, while achieving independent control of viewpoint, time, and illumination.
- Building rome in a day. ACM Communications, 2011.
- Factorized and controllable neural re-rendering of outdoor scene for photo extrapolation. In ACM MM, 2022.
- Efficient geometry-aware 3D generative adversarial networks. In CVPR, 2021.
- Depth synthesis and local warps for plausible image-based navigation. ACM TOG, 2013.
- Encoder-decoder with atrous separable convolution for semantic image segmentation. In ECCV, 2018.
- Hallucinated neural radiance fields in the wild. In CVPR, 2022.
- Unstructured light fields. In Eurographics, 2012.
- Fast dynamic radiance fields with time-aware neural voxels. SIGGRAPH Asia, 2022.
- Deepstereo: Learning to predict new views from the world’s imagery. In CVPR, June 2016.
- Building rome on a cloudless day. In ECCV, 2010.
- Towards internet-scale multi-view stereo. In CVPR, 2010.
- Learning neural volumetric representations of dynamic humans in minutes. In CVPR, 2023.
- Multi-view stereo for community photo collections. In ICCV, 2007.
- The lumigraph. In SIGGRAPH, 1996.
- Gancraft: Unsupervised 3d neural rendering of minecraft worlds. In ICCV, 2021.
- Disentangling random and cyclic effects in time-lapse sequences. ACM TOG, 2022.
- Learning-based view synthesis for light field cameras. ACM TOG, 2016.
- Light field rendering. In SIGGRAPH, 1996.
- Neural 3d video synthesis from multi-view video. In CVPR, 2022.
- Neural scene flow fields for space-time view synthesis of dynamic scenes. In CVPR, 2021.
- Learning intrinsic image decomposition from watching the world. In CVPR, 2018.
- Dynibar: Neural dynamic image-based rendering. In CVPR, 2023.
- Crowdsampling the plenoptic function. In ECCV, 2020.
- Efficient neural radiance fields for interactive free-viewpoint video. In SIGGRAPH Asia, 2022.
- Learning to factorize and relight a city. In European Conference on Computer Vision, pages 544–561. Springer, 2020.
- Neural sparse voxel fields. NeurIPS, 2020.
- 3d time-lapse reconstruction from internet photos. In ICCV, 2015.
- Time-lapse mining from internet photos. ACM TOG, 2015.
- NeRF in the Wild: Neural Radiance Fields for Unconstrained Photo Collections. In CVPR, 2021.
- Scene chronology. In ECCV, 2014.
- Occupancy networks: Learning 3d reconstruction in function space. In CVPR, 2019.
- Neural rerendering in the wild. In CVPR, 2019.
- Local light field fusion: Practical view synthesis with prescriptive sampling guidelines. ACM TOG, 2019.
- Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020.
- Differentiable volumetric rendering: Learning implicit 3d representations without 3d supervision. In CVPR, 2020.
- Deepsdf: Learning continuous signed distance functions for shape representation. In CVPR, 2019.
- Nerfies: Deformable neural radiance fields. In ICCV, 2021.
- Hypernerf: A higher-dimensional representation for topologically varying neural radiance fields. arXiv, 2021.
- Representing volumetric videos as dynamic mlp maps. In CVPR, 2023.
- D-nerf: Neural radiance fields for dynamic scenes. In CVPR, 2021.
- Beyond periodicity: Towards a unifying framework for activations in coordinate-mlps. In ECCV, 2022.
- Nerf for outdoor scene relighting. In ECCV, 2022.
- K-planes: Explicit radiance fields in space, time, and appearance. In CVPR, 2023.
- Multi-view matching for unordered image sets, or “how do i organize my holiday snaps?”. In ECCV, 2002.
- Probabilistic temporal inference on reconstructed 3d scenes. In CVPR, 2010.
- Structure-from-motion revisited. In CVPR, 2016.
- Pixelwise view selection for unstructured multi-view stereo. In ECCV, 2016.
- Shuhan Shen. Accurate multiple view 3d reconstruction using patch-based stereo for large-scale scenes. TIP, 2013.
- Implicit neural representations with periodic activation functions. NeurIPS, 2020.
- Deepvoxels: Learning persistent 3d feature embeddings. In CVPR, 2019.
- Photo tourism: exploring photo collections in 3d. In SIGGRAPH, 2006.
- Modeling the world from internet photo collections. IJCV, 2008.
- Neural 3D reconstruction in the wild. In SIGGRAPH, 2022.
- Factored time-lapse video. In ACM SIGGRAPH, pages 101–es. 2007.
- Single-view view synthesis with multiplane images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 551–560, 2020.
- Neural trajectory fields for dynamic novel view synthesis. arXiv, 2021.
- Nex: Real-time view synthesis with neural basis expansion. In CVPR, 2021.
- Space-time neural irradiance fields for free-viewpoint video. In CVPR, 2021.
- Editable free-viewpoint video using a layered neural representation. ACM TOG, 2021.
- High-quality video view interpolation using a layered representation. ACM TOG, 2004.