Bi-Grid Voxelization for Panoramic 3D Occupancy
- Bi-Grid Voxelization (BGV) is a dual-grid 3D modeling technique that uses parallel Cartesian and cylindrical-polar voxel grids to capture detailed near-field and uniform panoramic features.
- The method mitigates quantization biases by fusing multi-scale geometric features and aligning voxel grids to balance local precision with global angular continuity.
- BGV integrates into the OneOcc pipeline, enabling full-surround semantic scene completion crucial for legged and humanoid robotic navigation.
Bi-Grid Voxelization (BGV) is a 3D representation and feature-lifting technique developed for panoramic semantic occupancy prediction, specifically addressing quantization biases caused by the conflicting demands of dense near-field capture and uniform angular sampling required in 360° robot perception. BGV constructs and processes both a standard Cartesian and a cylindrical-polar voxel grid in parallel at multiple scales, allowing the system to capture high-fidelity spatial structures and uniform scene coverage simultaneously. This approach is integral to the OneOcc system for semantic scene completion with panoramic vision, enabling precise, full-surround environment modeling for legged and humanoid robots.
1. Motivation and Problem Setting
Robots with panoramic cameras—especially legged or humanoid platforms—require occupancy prediction systems capable of robustly modeling 3D free and occupied space for navigation. Equirectangular panoramas present a unique sampling dilemma: Cartesian voxel grids, though optimal for uniform spatial (x, y, z) capture near the robot, misalign with the annular, 360° camera field, leading to non-uniform sampling (over-sampling near field, under-sampling in far field) and artifacts at the azimuthal boundaries. Cylindrical-polar discretizations, defined in (r, φ, z) coordinates, maintain perfect angular continuity and uniformity across the panorama but sacrifice radial (near-field) detail critical for small obstacle detection and precise foot placement.
Relying solely on one grid structure necessitates a trade-off: either accept coarse far-field resolution or inadequate near-field representation, resulting in jagged, aliased free/occupied boundaries and diminished task performance. BGV is designed to mitigate these issues by unifying the strengths of both grid types—near-field precision and global angular uniformity—through dual-voxelization and cross-stream feature fusion.
2. Coordinate Systems and Transformations
BGV operates simultaneously in two coordinate systems:
- Cartesian (x, y, z): Right-handed world frame anchored to the robot. Suitable for uniform volume sampling and precise local geometry representation.
- Cylindrical-polar (r, φ, z): Where , , and . This system aligns with the panoramic image's inherent (φ) structure and models radial distance explicitly.
Coordinate transformations are defined as: and inverse: This dual system allows both local and angular correspondence between 3D world space and panoramic 2D feature representations.
3. Dual Voxel Grid Construction
Each input is lifted into two parallel 3D voxel grids per feature scale:
Cartesian voxel grid
- Bounds: , ,
- Indices: , ,
- Centroids: , and similarly for ,
Cylindrical-polar voxel grid
- Bounds: , (typically , ), identical z-bounds
- Indices: , , as above
- Centroids mapped to world frame: , with ,
Both grid types are constructed at multiple scales (e.g., downsampled by {1, 4, 8, 16} relative to input resolution).
4. Feature Lifting, View2View Sampling, and Fusion
At each scale , BGV operates as follows:
- Input acquisition: Two panoramic feature maps from the Dual-Projection fusion encoder: (equirectangular) and . These may be spatially warped by learned Gait Displacement Compensation offsets .
- Projection and sampling: For each voxel centroid (in both grids), project to each view and sample by 2D bilinear interpolation at : and analogue for
- Multi-scale fusion: Per-voxel, fuse across scales with convex weights (sum to 1 over s):
- Context injection: For each Cartesian voxel at level , its nearest neighbors in the polar grid are precomputed as index set . Polar features for these neighbors are aligned with a convolution and concatenated channel-wise with Cartesian features:
Pseudocode:
1 2 3 4 5 6 7 8 9 10 |
for scale ℓ in {1,4,8,16}: for each voxel c in Ca_grid_ℓ: p = project(c) + Δ_ℓ Vca_ℓ[c] = bilinear_sample(F_{1/ℓ}, p) for each voxel c in Po_grid_ℓ: p = project(c) + Δ_ℓ Vpo_ℓ[c] = bilinear_sample(F_{1/ℓ}, p) Vca[c] = Σ_ℓ α^{Ca_ℓ}(c) · Vca_ℓ[c] Vpo[c] = Σ_ℓ α^{Po_ℓ}(c) · Vpo_ℓ[c] Ca_fused_ℓ = Concat( Align( Vpo_ℓ[J_ℓ] ), Vca_ℓ ) |
5. Boundary Sharpening and Bias Mitigation
BGV achieves sharper free/occupied boundaries and reduces aliasing through complementary sampling densities:
- Near-field: Cartesian grid supplies fine, uniform spatial resolution, supporting accurate detection of sharp transitions such as contact edges or obstacles immediately adjacent to the robot.
- Far-field and angular continuity: Cylindrical grid maintains uniform sampling in azimuth () and models radial distance directly, capturing elongated or circumferential structures without under-sampling edge regions.
Dual-injection—concatenation of aligned polar features into the Cartesian processing stream at every decoder stage—allows the model to specialize, dynamically leveraging the strengths of each grid depending on scene context. Importantly, BGV does not require additional boundary-aware losses; the sharpening is an emergent property of complementary grid samplings and convex multi-scale fusion. Boundary-aware mechanisms external to BGV (such as the AMoE-3D decoder’s gradient-energy gate) are handled at later stages, with BGV itself purely geometry- and sampling-driven.
6. Integration Within OneOcc Semantic Occupancy Pipeline
BGV is interposed in the OneOcc network as follows:
- Inputs: Multi-scale 2D features from dual-projection encoders (), plus learned 2D motion warps from Gait Displacement Compensation.
- Outputs: Fused 3D Cartesian volumes with injected polar context () per decoder level, and optionally standalone polar 3D volumes for downstream use.
- Flow of information: Each volume feeds into the corresponding Hierarchical AMoE-3D decoder stage, where channel/spatial attention and MoE fusion integrate multi-scale context. Final semantic occupancy prediction is performed by a 1×1×1 classification head on the fully fused volume.
The end-to-end sequence is summarized as: Panorama → unwrapping → dual 2D encoders → GDC warps → BGV (dual 3D grid lifting + cross-injection) → AMoE-3D fusion → segmentation head → discretized semantic occupancy.
This procedure allows BGV to systematically exploit both local and panoramic scene geometry, yielding aliasing-robust, high-fidelity 3D occupancy outputs without the need for specialized edge-awareness outside standard volumetric segmentation procedures.
7. Significance and Context
By jointly modeling Cartesian and cylindrical-polar discretizations, BGV enables legged or humanoid robots with panoramic monocular vision to obtain reliable 3D semantic occupancy in omnidirectional, jitter-prone environments, where single-grid voxelization frameworks would suffer from discretization bias and boundary artifacts. The BGV approach is distinct in managing spatial-angular trade-offs by explicit grid-wise injection and scale-wise convex fusion, which together account for both near-robot accuracy and holistic 360° scene continuity. This suggests broader relevance for robotics researchers aiming to unify multi-view geometry, spatial perception, and efficient volumetric prediction in full-surround contexts.