Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

3D-Aware Scene Manipulation via Inverse Graphics (1808.09351v4)

Published 28 Aug 2018 in cs.CV, cs.GR, and eess.IV

Abstract: We aim to obtain an interpretable, expressive, and disentangled scene representation that contains comprehensive structural and textural information for each object. Previous scene representations learned by neural networks are often uninterpretable, limited to a single object, or lacking 3D knowledge. In this work, we propose 3D scene de-rendering networks (3D-SDN) to address the above issues by integrating disentangled representations for semantics, geometry, and appearance into a deep generative model. Our scene encoder performs inverse graphics, translating a scene into a structured object-wise representation. Our decoder has two components: a differentiable shape renderer and a neural texture generator. The disentanglement of semantics, geometry, and appearance supports 3D-aware scene manipulation, e.g., rotating and moving objects freely while keeping the consistent shape and texture, and changing the object appearance without affecting its shape. Experiments demonstrate that our editing scheme based on 3D-SDN is superior to its 2D counterpart.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Shunyu Yao (72 papers)
  2. Tzu Ming Harry Hsu (2 papers)
  3. Jun-Yan Zhu (80 papers)
  4. Jiajun Wu (249 papers)
  5. Antonio Torralba (178 papers)
  6. William T. Freeman (114 papers)
  7. Joshua B. Tenenbaum (257 papers)
Citations (84)

Summary

We haven't generated a summary for this paper yet.