Papers
Topics
Authors
Recent
2000 character limit reached

3D Gaussian Splatting (3D-GS)

Updated 7 December 2025
  • 3D Gaussian Splatting is a point-based 3D representation that models complex geometry using anisotropic Gaussian primitives with learnable geometric, radiometric, and semantic attributes.
  • The framework employs differentiable rasterization and advanced optimization to enable real-time novel view synthesis, robust 3D reconstruction, and effective compression.
  • Innovative methods like hierarchical densification and sub-vector quantization ensure high-fidelity rendering, significant storage reduction, and scalable performance.

3D Gaussian Splatting (3D-GS) is an explicit point-based 3D scene representation and rendering framework that models complex geometry and appearance using a collection of anisotropic Gaussian primitives with learnable geometric, radiometric, and sometimes semantic attributes. By leveraging the properties of Gaussian functions and corresponding differentiable rasterization techniques, 3D-GS serves as a foundation for real-time novel view synthesis, 3D reconstruction, compression, and interactive editing in both static and dynamic environments.

1. Mathematical Foundations and Rendering Pipeline

Each 3D-GS primitive is characterized by a mean position μR3\mu \in \mathbb{R}^3, a positive definite covariance matrix ΣR3×3\Sigma \in \mathbb{R}^{3 \times 3} encoding anisotropic spatial extent, an opacity or radiance weight w[0,1]w \in [0,1], and, typically, view-dependent color coefficients parameterized via low-order real spherical harmonics (SH) hRdh \in \mathbb{R}^d (with dd depending on SH degree). The Gaussian kernel for primitive ii is:

Gi(x)=wiexp(12(xμi)TΣi1(xμi))G_i(x) = w_i \exp\left(-\frac{1}{2} (x-\mu_i)^T \Sigma_i^{-1} (x-\mu_i)\right)

Rendering operates by projecting each 3D ellipsoid to a 2D ellipse on the image plane using an affine camera warp WW and local Jacobian JJ, giving projected mean μi\mu'_i and covariance Σi\Sigma'_i. The per-pixel contribution is

αi(x)=wiexp(12(xμi)T(Σi)1(xμi))\alpha_i(x) = w_i \exp\left(-\frac{1}{2} (x - \mu'_i)^T (\Sigma'_i)^{-1} (x - \mu'_i)\right)

Colors are composited via a front-to-back alpha-blending:

C(x)=k=1Kckαk(x)j<k(1αj(x))C(x) = \sum_{k=1}^K c_k\, \alpha_k(x)\prod_{j<k} (1 - \alpha_j(x))

where ckc_k is computed from the SH coefficients at the view direction for each primitive (Chen et al., 8 Jan 2024, Bao et al., 24 Jul 2024, Matias et al., 20 Oct 2025).

Tile-based rasterization assigns each projected ellipse to all overlapping image tiles; within each tile, splats are sorted by depth and blended efficiently in parallel. The full splat pipeline is differentiable, supporting gradient-based optimization for all parameters.

2. Optimization Strategies and Model Compression

Training typically begins by initializing primitive positions from a Structure-from-Motion (SfM) point cloud or similar sparse geometry, with random or heuristically chosen covariances. Photometric losses, often L1L_1 or L2L_2 plus perceptual terms like DSSIM, drive reconstruction accuracy across all training views. Regularization (e.g., penalizing degenerate or oversized covariances, enforcing sparsity in opacities, or depth-normal consistency) stabilizes geometric structure.

Practical deployment and memory efficiency require aggressive compression strategies:

  • Minimal Primitives via Local Distinctiveness: OMG (Lee et al., 21 Mar 2025) introduces a composite importance score: the base importance Iˉi\bar{I}_i is computed from blending weights across rays, modulated by a local distinctiveness term DiD_i derived from L1 color distances to spatially adjacent neighbors (found via Morton-order). Only Gaussians with the highest aggregate importance are retained, enabling up to 50% storage reduction with little loss in quality.
  • Attribute Compression: Instead of storing full per-primitive SH vectors, compact representations use short (R3\mathbb{R}^3) static and view-dependent features, fused with a small neural field (MLP) based on spatial encoding, which yields final SH coefficients and opacities.
  • Sub-Vector Quantization (SVQ): Features (e.g., (T,V), scale, rotation) are partitioned into subvectors quantized via small codebooks and trained using photometric loss. OMG achieves codebook sizes small enough for negligible overhead while maintaining output fidelity.
  • Pruning and Sensitivity Metrics: Principled Uncertainty Pruning (PUP-3DGS (Hanson et al., 14 Jun 2024)) computes per-primitive removal impact via second-order Taylor expansion of the reconstruction error, using a Fisher-information block-diagonal approximation for the local Hessian. Gaussian blocks with the lowest sensitivity scores are pruned in a multi-round refine loop, with up to 88% pruning achieved and minimal degradation.

Empirical results demonstrate that these techniques together can compress full scenes to 3–7 MB, maintain or improve metrics over prior methods, and enable rendering at hundreds (even thousands) of frames per second (Lee et al., 21 Mar 2025, Hanson et al., 14 Jun 2024).

3. Advances in Rendering Algorithms

3.1 Weighted Sum Rendering (WSR)

Standard alpha-blending in 3DGS is non-commutative, requiring per-frame, per-pixel depth sorting, which is computationally intensive—especially on resource-constrained hardware. Weighted Sum Rendering (WSR (Hou et al., 24 Oct 2024)) replaces depth-sorted compositing with a fully commutative, order-independent blend:

C=cBwB+i=1Nciαiw(di)wB+i=1Nαiw(di)C = \frac{c_B w_B + \sum_{i=1}^N c_i\,\alpha_i\,w(d_i)}{w_B + \sum_{i=1}^N \alpha_i w(d_i)}

where w(di)w(d_i) is a depth- or occlusion-weighting function (e.g., direct, exponential, or linear-corrected), cB,wBc_B, w_B denote background color/weight. All per-pixel contributions are summed independently, entirely eliminating depth sort and associated runtime artifacts such as “popping.”

Mobile GPU benchmarks show a 1.23×\times throughput improvement and reduced memory by exploiting the commutative structure. Image quality remains comparable or degrades minimally, with PSNR/SSIM/LPIPS on standard datasets nearly matching sorted blending (Hou et al., 24 Oct 2024).

3.2 Ray-Traced Extensions

Classic 3DGS integrates only “baked-in” lighting via SH and cannot natively simulate rays for shadows or reflection. RaySplats (Byrski et al., 31 Jan 2025) develops a ray-tracing-based splatting system by analytically solving for ray–ellipsoid intersections and front-to-back composition along arbitrary rays. This enables full physically-based shading, including soft shadows, specular highlights, and hybrid integration with triangle meshes. While rendering speed falls compared to rasterization, RaySplats matches baseline image quality and supports advanced lighting effects (Byrski et al., 31 Jan 2025).

4. Initialization, Densification, and Surface Fidelity

Accurate initialization and dynamic control over primitive density are critical for high-fidelity and efficient scene representation. Recent strategies address these challenges:

  • Geometry-Guided Initialization: GDGS (Wang et al., 1 Jul 2025) and GS-Net (Zhang et al., 17 Sep 2024) employ (supervised) MLPs or feed-forward networks that map SfM point clouds (and, if available, normals and color) to predict not only primitive centers but anisotropic covariances. This ensures that primitives are initialized near actual surfaces, reducing floaters and accelerating convergence.
  • Adaptive Densification: Region-aware metrics, combining local density variance, image-space gradients, and subcell sparsity (as in GDGS), or KNN-based splitting (EasySplat (Gao et al., 2 Jan 2025)), direct Gaussian spawning preferentially into under-sampled or highly textured regions. GeoTexDensifier (Jiang et al., 22 Dec 2024) further augments this with texture and geometry priors, leveraging monocular depth ratio checks and normal-guided placement to filter and align splits with true geometry, validated through monocular depth priors (ZoeDepth).
  • Topology-Aware Densification: Topology-GS (Shen et al., 21 Dec 2024) introduces persistent homology via Local Persistent Voronoi Interpolation (LPVI), adding new points if and only if the insertion induces negligible topological change in the local α\alpha-complex, preserving both geometric and topological consistency.

Such strategies yield sharper, surface-aligned splats, denser coverage in structured regions, and robust suppression of floaters, with state-of-the-art metrics for PSNR, SSIM, and LPIPS (Wang et al., 1 Jul 2025, Zhang et al., 17 Sep 2024, Jiang et al., 22 Dec 2024, Shen et al., 21 Dec 2024).

5. Scene Compression and Storage-Efficient Architectures

The increasing scale of 3DGS models and their unstructured nature drive innovations in storage and bandwidth reduction:

  • Tri-Plane and Neural Attribute Compression: TC-GS (Wang et al., 26 Mar 2025) encodes Gaussian attributes via a tri-plane structure: each primitive’s attributes are interpolated from three 2D feature planes, then locally decoded—along with those of KK nearest neighbors and the contracted position—by a compact MLP to obtain a probabilistic (mean/std) attribute code. Adaptive wavelet losses focus training on high-frequency details progressively. This achieves order-of-magnitude compression (up to 100×\times over vanilla 3DGS, 14×\times over ScaffoldGS) at competitive or improved render quality.
  • Sub-Vector Quantization (SVQ): As in OMG (Lee et al., 21 Mar 2025), splitting high-dimensional attributes into small subvectors for independent codebook quantization increases tolerance to lossy compression and enhances scalability.
  • Patchwise and Entropy-Optimized Coding: Quantization-based compression methods, sometimes cascaded with patchwise vector-quantized codes, further reduce size. Integration with hardware-aware codebook training is identified as a future direction (Wang et al., 26 Mar 2025).

6. Applications and Specialized Extensions

3D Gaussian Splatting forms the core of diverse downstream tasks:

  • Novel View Synthesis: High-fidelity, real-time rendering in both static and dynamic scenes, robust to occlusion and wide field-of-view requirements. Real-time VR applications are now possible, with state-of-the-art head-mounted displays attaining 72+ FPS and negligible latency using compact representations and rasterization optimization (Tu et al., 15 May 2025).
  • Compression and Streaming: Efficient representation and all-in-one pipelines support virtual navigation and immersive streaming for large-scale scans, aerial imagery, SLAM, and scientific visualization (see distributed 3DGS (Han et al., 15 Sep 2025)).
  • Augmented and Underwater Reality: UW-GS (Wang et al., 2 Oct 2024) integrates distance-dependent attenuation, scattering-aware color models, and dynamic outlier masking for accurate underwater scene reconstruction.
  • Scalable/Hierarchical Architectures: Scale-GS (Yang et al., 29 Aug 2025) proposes hierarchical anchor-based organization, multi-scale splitting, and redundancy-filtering via mask-based pruning for dynamic, streaming scenes.
  • Topology and Structural Integrity: Topology-GS (Shen et al., 21 Dec 2024) enforces topological consistency during densification and rendering, maintaining structural integrity in low-curvature areas through persistent homology constraints in both geometry and rendered RGB space.
  • Proxy-Guided Occlusion Culling: Proxy-GS (Gao et al., 29 Sep 2025) leverages a lightweight proxy mesh for high-resolution geometry-aware culling and densification, drastically reducing per-frame rendering load in occlusion-dense scenes.

7. Limitations and Current Research Directions

Despite rapid advances, several open challenges and research frontiers persist:

  • Dynamic/4DGS: Extensions to dynamic and multiframe scenes remain nontrivial, with current compression and attribute schemes targeting static scenes. Promising lines include temporal scaffolds and frame-wise bidirectional masking (Lee et al., 21 Mar 2025, Yang et al., 29 Aug 2025).
  • Relighting and BRDF Decoupling: Most methods bake illumination into SH colors, limiting physically-based relighting; hybrid pipelines and learned BRDF embeddings are active topics (Nguyen et al., 18 Nov 2025, Matias et al., 20 Oct 2025).
  • Memory-Bandwidth and Hardware Constraints: Further gains via entropy-optimized coding (beyond Huffman), hardware-aware codebook learning, and deployment on edge and non-volatile memory devices remain under investigation (Lee et al., 21 Mar 2025, Wang et al., 26 Mar 2025).
  • Topological Consistency at Scale: Scaling persistent homology-driven regularization to city-scale or unstructured scenes is an area for further paper (Shen et al., 21 Dec 2024).
  • Generalization and Initialization-Free Pipelines: Transferable, plug-and-play modules like GS-Net (Zhang et al., 17 Sep 2024) reduce reliance on scene-specific preprocessing and unlock cross-scene generalizability.
  • Hybrid and Multi-Representation Models: Integration of mesh, SDF, and triplane feature representations in a consistent, unified 3DGS framework is a developing area, promising further efficiency and fidelity (Wang et al., 26 Mar 2025, Matias et al., 20 Oct 2025).

3D-GS thus remains an active research paradigm bridging point-based explicit graphics, neural rendering, and geometric learning, rapidly evolving to solve the demands of high-fidelity, real-time, scalable 3D representation across domains (Bao et al., 24 Jul 2024, Chen et al., 8 Jan 2024, Lee et al., 21 Mar 2025, Hou et al., 24 Oct 2024, Tu et al., 15 May 2025, Nguyen et al., 18 Nov 2025, Wang et al., 26 Mar 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
Slide Deck Streamline Icon: https://streamlinehq.com

Whiteboard

Forward Email Streamline Icon: https://streamlinehq.com

Follow Topic

Get notified by email when new papers are published related to 3D Gaussian Splatting (3D-GS).