Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
153 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Fillerbuster: Multi-View Scene Completion for Casual Captures (2502.05175v1)

Published 7 Feb 2025 in cs.CV and cs.GR

Abstract: We present Fillerbuster, a method that completes unknown regions of a 3D scene by utilizing a novel large-scale multi-view latent diffusion transformer. Casual captures are often sparse and miss surrounding content behind objects or above the scene. Existing methods are not suitable for handling this challenge as they focus on making the known pixels look good with sparse-view priors, or on creating the missing sides of objects from just one or two photos. In reality, we often have hundreds of input frames and want to complete areas that are missing and unobserved from the input frames. Additionally, the images often do not have known camera parameters. Our solution is to train a generative model that can consume a large context of input frames while generating unknown target views and recovering image poses when desired. We show results where we complete partial captures on two existing datasets. We also present an uncalibrated scene completion task where our unified model predicts both poses and creates new content. Our model is the first to predict many images and poses together for scene completion.

Summary

  • The paper introduces Fillerbuster, a novel latent diffusion transformer trained to complete multi-view 3D scenes from sparse, casual captures by generating unknown views conditioned on known images and poses.
  • Fillerbuster unifies content generation and pose recovery, enabling uncalibrated scene completion tasks where the model predicts plausible camera poses while creating missing scene parts.
  • Evaluated on existing datasets, Fillerbuster significantly improves scene completion quality over baselines, demonstrating higher PSNR, SSIM, and lower LPIPS.

The paper "Fillerbuster: Multi-View Scene Completion for Casual Captures" introduces a novel method, Fillerbuster, for completing unknown regions in 3D scenes captured casually using a large-scale multi-view latent diffusion transformer.

The key contributions of this paper are:

  • A generative model is trained to consume a large context of input frames while generating unknown target views and recovering image poses when desired. The model is unified to predict many images and poses together for scene completion.
  • The model leverages a latent diffusion transformer with a flow-matching loss for inpainting in latent space, conditioned on known images and camera poses (represented as raymaps).
  • The model is evaluated by completing partial captures on existing datasets and by introducing an uncalibrated scene completion task where the unified model predicts both poses and creates new content.

The challenges addressed include:

  • Casual captures are often sparse and miss surrounding content, making existing methods unsuitable as they focus on refining known pixels with sparse-view priors or completing object sides from limited photos.
  • The desire to complete areas missing and unobserved from hundreds of input frames, even without known camera parameters.

The technical approach involves:

  • A latent diffusion transformer model is proposed that denoises multiple input images and calibrated camera poses with masks indicating known and unknown regions.
  • The model handles a sequence of NN elements with images IiRH×W×3I_i \in \mathbb{R}^{H \times W \times 3} (where HH is height and WW is width), raymaps RiRH×W×6R_i \in \mathbb{R}^{H \times W \times 6} with origin and direction per pixel, valid image masks MiIRH×W\mathcal{M}^\text{I}_i \in \mathbb{R}^{H \times W}, and valid ray masks MiRRH×W\mathcal{M}^\text{R}_i \in \mathbb{R}^{H \times W}, with the objective to predict p(I,RIMI,RMR)p(I,R \mid I \odot \mathcal{M}^\text{I}, R \odot \mathcal{M}^\text{R}).
    • IiI_i: Images
    • HH: Height of image
    • WW: Width of image
    • RiR_i: Raymaps
    • MiI\mathcal{M}^\text{I}_i: Valid image masks
    • MiR\mathcal{M}^\text{R}_i: Valid ray masks
  • The architecture uses a DiT and trains with the flow matching objective. Separate VAEs are used for images and poses encoded as raymaps, with EI\mathcal{E}^\text{I} denoting the image encoder and ER\mathcal{E}^\text{R} denoting the raymap encoder.
    • EI\mathcal{E}^\text{I}: Image encoder
    • ER\mathcal{E}^\text{R}: Raymap encoder
  • The sequence is prepared as
    • si,ts_{i,t}: Sequence
    • z~i,tI\tilde{z}^\text{I}_{i,t}: Compressed latent image
    • D\mathcal{D}: Downscaling operation
    • z~i,tR\tilde{z}^\text{R}_{i,t}: Compressed latent raymap
  • Two forms of positional embeddings are used: 2D layout embeddings and index embeddings, to handle varying sequence lengths.

The results include:

  • Demonstration of completing casual captures from the Nerfbusters dataset, showing improved consistency compared to baselines like 3DGS and a CAT3D-sized conditioning model.
  • Introduction of the "uncalibrated scene completion" task, generating fly-throughs from sparse unposed photos, predicting plausible camera poses, and completing missing content.
  • Outperformance of NeRFiller on the NeRFiller dataset in terms of quality and consistency, measured by PSNR, SSIM, and LPIPS metrics. PSNR increased from 25.57 to 29.60, SSIM from 0.89 to 0.92, and LPIPS decreased from 0.182 to 0.096.
    • PSNR: Peak Signal-to-Noise Ratio
    • SSIM: Structural Similarity Index Measure
    • LPIPS: Learned Perceptual Image Patch Similarity
  • Ablation studies validate the importance of index embeddings and pose prediction for model performance.

The method shows improvements in completing casual captures, uncalibrated scene completion, and multi-view inpainting tasks. The unified model's ability to jointly complete the unobserved content and recover poses is valuable for casual scene completion.