Papers
Topics
Authors
Recent
Search
2000 character limit reached

Unifi3D: A Study on 3D Representations for Generation and Reconstruction in a Common Framework

Published 2 Sep 2025 in cs.GR, cs.CV, and cs.LG | (2509.02474v1)

Abstract: Following rapid advancements in text and image generation, research has increasingly shifted towards 3D generation. Unlike the well-established pixel-based representation in images, 3D representations remain diverse and fragmented, encompassing a wide variety of approaches such as voxel grids, neural radiance fields, signed distance functions, point clouds, or octrees, each offering distinct advantages and limitations. In this work, we present a unified evaluation framework designed to assess the performance of 3D representations in reconstruction and generation. We compare these representations based on multiple criteria: quality, computational efficiency, and generalization performance. Beyond standard model benchmarking, our experiments aim to derive best practices over all steps involved in the 3D generation pipeline, including preprocessing, mesh reconstruction, compression with autoencoders, and generation. Our findings highlight that reconstruction errors significantly impact overall performance, underscoring the need to evaluate generation and reconstruction jointly. We provide insights that can inform the selection of suitable 3D models for various applications, facilitating the development of more robust and application-specific solutions in 3D generation. The code for our framework is available at https://github.com/isl-org/unifi3d.

Summary

  • The paper introduces a unified benchmarking framework for 3D representations by isolating variables like model architecture and preprocessing.
  • It quantitatively evaluates generation and reconstruction performance across methods such as voxel grids, SDFs, and NeRFs using metrics like Chamfer Distance and F-score.
  • Experimental results highlight trade-offs, with DualOctree excelling in generation metrics and SDFs achieving superior reconstruction fidelity.

Unifi3D: Unified Benchmarking of 3D Representations for Generation and Reconstruction

Introduction

The paper "Unifi3D: A Study on 3D Representations for Generation and Reconstruction in a Common Framework" (2509.02474) presents a comprehensive, modular benchmarking framework for evaluating 3D representations in both generative and reconstructive tasks. Unlike the standardized pixel-based representations in 2D image synthesis, 3D generative modeling remains fragmented, with diverse representations such as voxel grids, signed distance functions (SDFs), point clouds, octrees, triplanes, and neural radiance fields (NeRFs). The Unifi3D framework enables controlled, quantitative comparison of these representations by integrating them into a unified pipeline, isolating the effects of representation from confounding factors such as model architecture and preprocessing. Figure 1

Figure 1: Overview of the steps involved in a standard 3D generation pipeline: mesh conversion, encoder compression, latent diffusion, decoder reconstruction, and mesh post-processing.

Unified Pipeline and Representations

The pipeline consists of four stages: mesh-to-representation conversion, representation compression via autoencoders (AE, VAE, VQ-VAE), latent generation using diffusion models (DiT, U-Net), and mesh reconstruction (e.g., Marching Cubes). The framework supports plug-and-play integration of multiple representations, including:

  • Voxel Grids: Dense or sparse occupancy grids, compatible with CNNs but memory-intensive.
  • SDF Grids: Scalar fields encoding signed distances, enabling smooth surface modeling and efficient grid-based processing.
  • Point Clouds: Sets of 3D points, often used as intermediate representations.
  • Dual Octree Graphs: Hierarchical, adaptive volumetric encoding using GNNs for efficient high-resolution geometry.
  • Triplanes: Three orthogonal 2D feature planes, balancing memory efficiency and spatial expressiveness.
  • NeRFs: Implicit neural fields for view synthesis, parameterized by MLPs.

Each representation is paired with a suitable encoder/decoder architecture and a generative model, ensuring fair comparison by controlling for model capacity and training protocol.

Evaluation Protocol

The framework advocates joint evaluation of reconstruction and generation, recognizing that reconstruction errors (from mesh conversion and compression) upper-bound generative performance. The protocol includes:

  1. Reconstruction Quality: Mesh →\rightarrow representation →\rightarrow mesh, measured by Chamfer Distance (CD), F-score, and Normal Consistency (NC).
  2. Compression Performance: Mesh →\rightarrow representation →\rightarrow latent →\rightarrow representation →\rightarrow mesh.
  3. Generalization: Out-of-distribution (OOD) reconstruction, e.g., encoder trained on one category, tested on another.
  4. Generation Metrics: Unconditional generation assessed by Coverage (COV), Minimum Matching Distance (MMD), and 1-Nearest Neighbor Accuracy (1-NNA), all based on CD between generated and reference meshes.
  5. Human Preference: User study with Bradley-Terry modeling to capture perceptual quality. Figure 2

Figure 2

Figure 2

Figure 2: Distributional metrics (COV, MMD, 1-NNA) for random subsets of ShapeNet airplane category, illustrating metric sensitivity to sample size and dataset bias.

Experimental Results

Generation Performance

The DualOctree VAE with U-Net diffusion achieves the best quantitative generation metrics (COV=0.365, MMD=0.031, 1-NNA=0.824), outperforming SDF, Shape2VecSet, Triplane, and Voxel representations. However, user studies rank SDF-generated meshes highest in perceptual quality, likely due to their smooth surfaces, while DualOctree tends to produce more complex but artifact-prone assets. Figure 3

Figure 3: Rankings of 3D representations based on generation and reconstruction metrics, including user study scores and surface-to-volume ratio.

Figure 4

Figure 4: Qualitative results for mesh generation across representations, highlighting differences in surface smoothness and complexity.

Reconstruction Performance

SDF and Voxel grid encodings with AE/VAEs yield the highest reconstruction fidelity (F-score ≈\approx 88%), with SDF AE achieving the best overall scores. NeRF encoding performs worst, attributed to architectural modifications for comparability. DualOctree achieves competitive results with minimal latent size, demonstrating the efficiency of adaptive hierarchical encoding. Figure 5

Figure 5: Qualitative results for mesh reconstruction, showing error modes such as loss of thin structures and surface artifacts.

Generalization and Scalability

Most encoders generalize well to OOD categories, except NeRF and DualOctree, which struggle due to overfitting and low latent dimensionality, respectively. The trade-off between latent size, reconstruction error, and runtime is evident: DualOctree achieves good accuracy with small latent size, while Voxel grids require larger latents for high fidelity. Figure 6

Figure 6: Complexity analysis, relating reconstruction quality to runtime and latent size.

Preprocessing and Error Decomposition

Mesh preprocessing (e.g., manifoldization) significantly affects reconstruction and generation. The proposed flood-fill SDF conversion avoids artificial thickening, preserving original geometry better than Mesh2SDF or ManifoldPlus. Reconstruction and compression errors constitute up to 39% of the total generative error, underscoring the necessity of joint evaluation. Figure 7

Figure 7: Comparison of mesh conversion methods in terms of Chamfer distance, demonstrating the impact of preprocessing on reconstruction error.

Figure 8

Figure 8: Effect of grid resolution on round-trip conversion errors, highlighting the trade-off between fidelity and computational cost.

Implications and Best Practices

The study establishes several best practices for 3D generative modeling:

  • Joint Evaluation: Report both reconstruction and generation errors, as reconstruction quality upper-bounds generative performance.
  • Direct Mesh Comparison: Use metrics computed on original meshes, avoiding distortions from preprocessing or derivative representations.
  • Sample Size: Use sufficiently large sample sets (>>200) for robust distributional metrics.
  • Human Evaluation: Incorporate user studies to capture perceptual quality not reflected in geometric metrics.

The modular, open-source codebase facilitates rapid prototyping and benchmarking of new representations and generative models, promoting reproducibility and fair comparison.

Limitations and Future Directions

The framework focuses on direct 3D generation, excluding optimization-based methods (e.g., Score Distillation Sampling) and Gaussian Splatting approaches, which are prevalent in text-to-3D and view synthesis but have different computational and output characteristics. Future work could extend the analysis to autoregressive models, additional datasets (e.g., ObjaverseXL), and more complex object categories (e.g., articulated or scene-level generation).

Open challenges remain in modeling interior structures, scaling to scene-level synthesis, and enabling controllable articulation. The presented pipeline provides a foundation for systematic exploration of these directions.

Conclusion

Unifi3D delivers a rigorous, unified framework for benchmarking 3D representations in generative and reconstructive tasks. The results demonstrate that representation choice critically affects both quantitative and perceptual outcomes, with SDF and DualOctree offering complementary strengths. The study's recommendations and open-source tools set a new standard for reproducible, transparent evaluation in 3D generative modeling, enabling informed selection and development of representations for diverse applications.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

GitHub