Papers
Topics
Authors
Recent
2000 character limit reached

VLM-based Physical 3D Generative Model

Updated 18 November 2025
  • VLM-based Physical 3D Generative Model is a framework that integrates vision-language cues with neural 3D generation to create simulatable, part-aware assets.
  • It employs efficient tokenization and hierarchical semantic extraction to generate high-fidelity meshes with explicit joint and physical annotations.
  • The model demonstrates improved simulation fidelity and physical parameter accuracy, supporting advanced applications in robotics and embodied AI.

A vision-LLM (VLM)-based physical 3D generative model is a system integrating large-scale multimodal LLMs with neural 3D generation architectures to synthesize simulatable, physically annotated, and part-aware 3D assets directly from visual and/or textual input. These models push generative 3D modeling beyond visual realism toward physical fidelity, semantic decomposition, and structural editability, directly supporting applications in robotics, simulation, and embodied AI.

1. Conceptual Foundations and Motivation

Traditional 3D generative models produce single-mesh assets optimized for visual appearance but lack explicit representations of object articulation, physical parameters, or functional part structure. This limits downstream editability, animation, and especially suitability for simulation or physical reasoning. VLM-based physical 3D generative models address this by tightly integrating neural generation (mesh, voxel, point cloud, or radiance field) with a vision-LLM that mediates semantic, geometric, and physical knowledge. The VLM acts as an interface for task-grounded decomposition, functional annotation, and geometry generation, enabling control and supervision by rich natural language and image prompts (Cao et al., 17 Nov 2025).

2. Key Architectural Components

The characteristic pipeline for a VLM-based physical 3D generative model, exemplified by PhysX-Anything (Cao et al., 17 Nov 2025), is composed of the following stages:

  1. Input Encoding & Semantic Extraction The input image is encoded via a multimodal VLM (e.g., Qwen2.5-VL). Multi-round “chat” or autoregressive prompting yields a global JSON-style representation describing the object’s absolute scale, part–subpart hierarchy, kinematic graph (joint types, axes, ranges), and material and affordance annotations.
  2. 3D Geometry Tokenization and Generation The input object is voxelized into a coarse 32³ occupancy grid. Occupied voxels are mapped to indices, compressed as merged index ranges to fit the VLM token budget (achieving ∼193× reduction versus naïve serialization). The VLM, conditioned on the physical description, autoregressively generates the geometry for each part independently as token ranges.
  3. Fine Geometry and Physical Decoding A controllable flow transformer and latent diffusion mesh decoder refine the coarse voxel representation into high-fidelity, watertight meshes. The part-wise geometry, material, joint, and other physical attributes are assembled in a URDF/XML output, suitable for direct simulation.
  4. Physical Simulation-Ready Export The resulting XML/URDF includes explicit joint definitions, mass/inertia tensors, and material assignments as well as mesh geometry for direct loading into physics engines (e.g., MuJoCo) (Cao et al., 17 Nov 2025).

The system avoids introducing specialized tokens during VLM fine-tuning and instead relies on efficient tokenization and structured prompting to maximize context utilization.

3. VLM Integration and Prompting Strategies

VLMs enable end-to-end semantic supervision and fine-grained control by processing user-supplied images and text, and generating interpretable physical descriptions. In PhysX-Anything, the VLM is fine-tuned to emit full part–physical trees and geometric tokenizations by jointly conditioning on the image and context (Cao et al., 17 Nov 2025). Prompting includes multi-round “chat” interactions to elicit explicit hierarchical JSON representations covering:

  • Object and part dimensions in metric units
  • Material class and prior (metal, plastic, etc.)
  • Part kinematics (joint type, axis, range)
  • Affordance or function descriptors

This approach obviates the need for per-category heuristics and aligns the generative process with both human intent and simulation requirements.

4. Physical Representation Modeling and Decoding

Every generated asset is parameterized by both geometric and physical properties:

  • Geometry: Produced as a nested occupancy grid, refined to a mesh per part. The tokenized representation allows the VLM to generate coarse shape which is then upsampled using a diffusion process and finally reconstructed into a mesh.
  • Physical Parameters: Each part is assigned a mass m=ρVdVm = \rho \int_V dV (with ρ\rho inferred from predicted material), inertia tensor I=Vρ(r)(r2I3rr)drI = \int_V \rho(r) (\|r\|^2 I_3 - r r^\top) dr, and joint parameters (type, axis, origin, bounds) output in standard URDF/XML fields. These enable direct simulation with correct dynamics and articulation.
  • Articulation and Joint Modeling: Part hierarchies support rigid, revolute, and prismatic joints, with explicit axes and limits derived from the VLM-prompted physical description (Cao et al., 17 Nov 2025).

5. Data, Learning Objectives, and Empirical Results

Dataset Construction: PhysX-Mobility, an extension of PartNet-Mobility, provides 2,315 objects with systematically annotated part hierarchies, materials, dimensions, joint definitions, and affordances covering 47 categories. Manual correction and physical prior assignment ensure dataset fidelity (Cao et al., 17 Nov 2025).

Loss Functions:

  • VLM cross-entropy: next-token prediction on tokenized geometry and physical fields.
  • Diffusion geometry loss: reconstruction loss in voxel grid (mean squared error).
  • Mesh reconstruction loss: Chamfer distance plus Eikonal regularization for smoothness.
  • Physics consistency loss: 1\ell_1 error on mass and material, Frobenius norm on inertia accuracy.
  • Articulation constraint: joint bounds and regularization.
  • Total loss combines these with empirically determined coefficients.

Experimental Evaluation:

  • On PhysX-Mobility, PhysX-Anything achieves higher PSNR (20.35 dB) and lower Chamfer distance (14.43) than competitive baselines, with marked improvements in scale prediction accuracy (0.3 m mean absolute error versus 43.44 m).
  • For material and affordance recognition, 17.5% and 14.3% accuracy are reported (vs. 6.3% and 9.8% for previous best, respectively).
  • User paper: 0.98 geometry preference versus 0.61 for the next best model; VLM judgment kinematics accuracy 0.94 (Cao et al., 17 Nov 2025).

Ablation: The compressed tokenization is critical: reducing geometry to 156 tokens preserves geometric detail as measured by PSNR and Chamfer distance while allowing the VLM to process explicit geometry in a single context window (Cao et al., 17 Nov 2025).

Simulation Validation: Output assets are imported into MuJoCo-style simulators for contact-rich manipulation policy learning. Policy reward curves converge comparably to those using hand-modeled ground-truth CAD, confirming that generated assets have physically plausible geometry, articulation, and physical parameters.

6. Limitations, Open Challenges, and Future Directions

Several challenges are highlighted:

  • Representational limits: The 32³ voxel grid bounds geometric fidelity on fine curvature. Higher-resolution (octree or implicit function) representations are suggested as remedies.
  • Physical parameter generalization: Currently, only basic joint types (rigid, revolute, prismatic) and single-material/part modeling are supported; more complex linkages and material heterogeneity remain open.
  • VLM limitations: Token budget constraints restrict the topological complexity of assets processable by the VLM in a single pass.
  • Simulation compatibility: While validated in MuJoCo, extensions to other engines (e.g., Bullet, Drake) and to non-rigid/deformable assets are necessary for broad adoption (Cao et al., 17 Nov 2025).
  • Scaling: Extension to Internet-scale weakly-supervised pretraining on image–3D pairs and better annotation pipelines (potentially via improved scene decomposition models) are proposed as future research avenues.

Prior physically grounded 3D generation approaches (e.g., PhysXGen (Cao et al., 16 Jul 2025)) rely on separate annotation models or human-in-the-loop pipelines for physical labels, often decoupling geometry and physical property generation. PhysX-Anything unifies these via end-to-end VLM-based learning which directly outputs simulation-ready structure in conjunction with geometry. Alternative pipelines such as MMPart (Bonakdar et al., 20 Sep 2025) focus on part-aware segmentation and multi-view 3D reconstruction but do not output explicit jointed or physically simulatable assets. Score Distillation Sampling-based approaches (e.g., VLM3D (Bai et al., 19 Sep 2025)) introduce VLMs as differentiable reward functions to guide 3D consistency and semantic fidelity during neural 3D asset optimization but lack explicit grounding of mechanical properties and joint articulation.

Approach Key Output Physical Parameters Part/Joint Structure Input Modality
PhysX-Anything Sim-ready mesh + URDF/XML Yes Yes In-the-wild image
PhysXGen Structured mesh + annotations Yes Partial 3D asset
MMPart Part-aware mesh No Partial Single image + list
VLM3D (SDS) Neural 3D (NeRF/splat/mesh) No No Text prompt

PhysX-Anything is unique in integrating VLM-based explicit semantic understanding, part decomposition, efficient geometry tokenization, and articulated simulation-ready asset output in a single framework (Cao et al., 17 Nov 2025).


References

PhysX-Anything: Simulation-Ready Physical 3D Assets from Single Image (Cao et al., 17 Nov 2025) PhysX-3D: Physical-Grounded 3D Asset Generation (Cao et al., 16 Jul 2025) MMPart: Harnessing Multi-Modal LLMs for Part-Aware 3D Generation (Bonakdar et al., 20 Sep 2025) Vision-LLMs as Differentiable Semantic and Spatial Rewards for Text-to-3D Generation (Bai et al., 19 Sep 2025)

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to VLM-based Physical 3D Generative Model.