Papers
Topics
Authors
Recent
2000 character limit reached

MeshRipple: Autoregressive Mesh Generation

Updated 16 December 2025
  • MeshRipple is a structured autoregressive framework that generates 3D triangle meshes through a frontier expansion process analogous to ripple propagation.
  • It employs expansive prediction with candidate masking and a dedicated transformer head to enforce connectivity and surface integrity.
  • It integrates a sparse-attention dual-memory mechanism to capture local and global dependencies, achieving state-of-the-art fidelity and geometric performance.

MeshRipple refers to a structured autoregressive framework for 3D triangle mesh generation that expands surfaces from an active frontier in a manner analogous to a ripple propagating over a surface. The approach integrates a frontier-aware breadth-first search (BFS) tokenization, an expansive prediction mechanism responsible for coherent connectivity, and a sparse-attention dual-memory architecture to capture both local and global mesh dependencies. MeshRipple achieves state-of-the-art fidelity and topological completeness in both dense and artistically stylized mesh synthesis, outperforming recent autoregressive models across standard geometric benchmarks (Lin et al., 8 Dec 2025).

1. Formal Specification and Tokenization

Meshes are represented as M=(V,F)M = (V, F), where VV is a set of quantized vertex coordinates and FF is a sequence of oriented faces (typically triangles with counterclockwise ordering). MeshRipple serializes faces into an ordered sequence via a half-edge-based BFS. The procedure begins with a seed face F0F_0, maintaining a frontier queue BB of faces adjacent to the active boundary. For each dequeued face FpF_p, its three outgoing half-edges (in fixed CCW order) are examined; any unvisited adjacent face FiF_i is marked, enqueued, and appended to the output serialization SS.

For each newly serialized face FiF_i, a root index rir_i is stored, indicating the frontier face from which FiF_i was reached. The set of active frontier faces BiB_i at step ii consists of those in S[1i]S[1{\ldots}i] with unvisited neighbors. This tokenization guarantees that the neighborhood required for predicting Fi+1F_{i+1} lies within BiB_i, which typically resides near the tail of SS. Thus, truncated windows of the last LL tokens generally suffice for local context. Each face token is uniquely determined by

Fi{ri,Bi}F_i \leftarrow \{r_i, B_i\}

plus its vertex triple (Lin et al., 8 Dec 2025).

2. Expansive Prediction and Connectivity Enforcement

MeshRipple incorporates a joint prediction strategy to enforce surface connectedness and mesh integrity. At each generation step, the model predicts:

  • The triangle Fi+1F_{i+1} to attach to the current root face FrF_r,
  • The root offset Δi+1\Delta_{i+1}, defining the position of the next root pointer within BiB_i.

Mathematically, let rir_i and ri+1r_{i+1} denote current and next root indices, so Δi+1=ri+1ri\Delta_{i+1} = r_{i+1} - r_i. The model factorizes the conditional as

p(Fi+1,Δi+1S[1i])=p(Fi+1S[1i])p(Δi+1S[1i])p(F_{i+1}, \Delta_{i+1} \mid S[1{\ldots}i]) = p(F_{i+1} \mid S[1{\ldots}i]) \cdot p(\Delta_{i+1} \mid S[1{\ldots}i])

where the root-offset term is predicted by a dedicated head on a mid-level transformer layer. During inference, the set of candidate triangle tokens is masked to include only those sharing a half-edge with the current root FrF_r, strictly enforcing surface attachment and eliminating fragmented or disconnected span. Training minimizes combined cross-entropy over both outputs: L=k[logpθ(Fk)+logpθ(Δk)]\mathcal{L} = - \sum_k \Big[\log p_\theta(F_k| \ldots ) + \log p_\theta(\Delta_k | \ldots )\Big] (Lin et al., 8 Dec 2025).

3. Sparse Global Memory via NSCA

Standard sliding-window architectures impose hard limits on receptive field size, impeding the capture of long-range geometric dependencies and symmetry. MeshRipple addresses this by introducing Non-overlapping Sparse Compressed Attention (NSCA), which hierarchically compresses all but the most recent LL tokens:

  1. The serialized sequence SS is partitioned into non-overlapping blocks of size BfB_f.
  2. Each block is summarized by a learned MLP producing compressed key-value pairs.
  3. At each decoding step:
    • Block-level scoring identifies the kk most relevant blocks via query-key dot-products,
    • For these blocks, the original token sequences are retrieved and attended at finer granularity,
    • Local attention over the immediate window proceeds concurrently.

Causal masking ensures strictly autoregressive access; tokens and blocks beyond the current generation index are unobservable. This architecture provides an effectively unbounded receptive field while keeping per-token time/memory complexity sub-quadratic: O(L2+S/Bf+kBf)O(L^2 + |S|/B_f + k B_f). Setting BfLB_f \gg L and kS/Bfk \ll |S|/B_f prevents out-of-memory failures encountered in dense attention (Lin et al., 8 Dec 2025).

4. Generation Loop and Inference Dynamics

The core MeshRipple autoregressive loop involves:

  • Initializing SgeneratedS_{\rm generated} with a start token and the frontier BfrontierB_{\rm frontier} with the seed face,
  • At each step:
    • Preparing the AR input window WW (last LL faces),
    • Computing hidden states using Transformer blocks with custom attention masks,
    • Predicting the next face Fi+1F_{i+1} with candidate masking for admissible attachments,
    • Predicting the root offset Δi+1\Delta_{i+1},
    • Updating sequence, frontier, and root pointer via current_root_indexcurrent_root_index+Δi+1current\_root\_index \leftarrow current\_root\_index + \Delta_{i+1},
    • Iterating until a stopping criterion is met.

Out-of-window context is accessible through NSCA, while the explicit frontier and attachment constraint guarantee the mesh surface grows as a single expanding component (Lin et al., 8 Dec 2025).

5. Empirical Performance and Ablations

MeshRipple has been evaluated against leading mesh autoregressive generators on both dense-mesh and artist-mesh datasets. Key geometric metrics include Chamfer distance (CD), Hausdorff distance (HD), and Normal Consistency (NC):

Model Chamfer (×10⁻³) ↓ Hausdorff ↓ Normal Consistency ↑
MeshAnythingV2 109.64 0.2314 –0.0096
BPT 60.19 0.1089 0.6066
DeepMesh 50.27 0.0893 0.6025
MeshRipple 48.73 0.1057 0.6280

On artist-mesh benchmarks, MeshRipple also outperforms FastMesh and BPT, with lowest CD and HD and highest NC (Lin et al., 8 Dec 2025). Qualitative assessments indicate that MeshRipple avoids the large holes and disconnected fragments observed in fixed-order AR baselines.

Ablation studies demonstrate:

  • Removing the frontier mask increases CD by +2.12 and HD by +0.0141,
  • Removing context injection increases CD by +7.89, HD by +0.0118,
  • Removing the root constraint sharply increases CD (+12.41) and HD (+0.0122),
  • Omitting NSCA only slightly affects CD (–0.24) but induces significant resource overhead.

6. Limitations and Prospects for Extension

MeshRipple can be sensitive to label noise and to meshes with extreme non-manifold topologies; in such regimes prediction coherence may degrade. The NSCA architecture in principle enables scaling beyond 100k faces, contingent on further increases in input quantization granularity and block sizing. Adaptation to quadrilateral (quad) and hexahedral mesh structures is a natural extension. The methodology is also compatible with reinforcement-learning fine-tuning (DPO/RL-HF) for mesh stylization as demonstrated in related works such as DeepMesh and MeshRFT.

A plausible implication is that the explicit frontier-tracking mechanism, when coupled with global memory retrieval, may serve as a generalizable approach for structured autoregressive generation in other domains with complex topological dependencies (Lin et al., 8 Dec 2025).

7. Comparison with Dynamic Mesh Approaches in Physics

The term “MeshRipple” has also been used in the context of numerically simulating ripple morphodynamics due to colloidal deposition in flowing fluids (Hewett et al., 2018). In Hewett & Sellier’s work, the arbitrary Lagrangian–Eulerian (ALE) approach is used to couple fluid flows with evolving boundary geometries by tracking mesh nodes and enforcing mesh quality via smoothing and shuffle algorithms. The similarity to the generative MeshRipple approach is nominal; in physics, “ripple” refers to evolving physical undulations on a material interface, modeled via dynamic mesh deformation and interface velocity determined by particle flux. In contrast, in generative modeling, “MeshRipple” refers to a deterministic surface traversal and completion process inspired by ripple propagation.

Both contexts, however, address the preservation of surface connectivity and mesh quality—either physically, via dynamic node management, or probabilistically, via autoregressive growth with enforced attachment constraints (Hewett et al., 2018, Lin et al., 8 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to MeshRipple.