AutoBrep: End-to-End Neural B-Rep Generation
- AutoBrep is a model that automatically synthesizes watertight B-Rep solid models by unifying parametric geometry and discrete topology in a single sequence.
- It employs a decoder-only Transformer with unified tokenization for both geometric and topological structures, reducing error propagation and enhancing inference efficiency.
- Empirical results demonstrate improved validity, coverage, and speed, establishing AutoBrep as a scalable solution for high-fidelity CAD modeling.
AutoBrep refers to a class of models and methodologies for fully automatic, end-to-end generation of boundary representation (B-Rep) solid models in Computer-Aided Design (CAD), synthesizing both precise geometry and watertight topology directly from learned data-driven representations. AutoBrep approaches—exemplified by the "AutoBrep" autoregressive Transformer model—address the longstanding challenge of direct B-Rep synthesis by unifying the generation of parametric geometry (NURBS, B-splines, etc.) and discrete topology (face adjacency, edge loops) within a single, coherent sequential model (Xu et al., 2 Dec 2025). This paradigm replaces prior decomposed, multi-stage B-Rep pipelines, reducing error propagation and enabling scalable, high-fidelity CAD modeling at industrial complexity.
1. Motivation and Background
Boundary representation (B-Rep) is the canonical data structure for solid modeling in CAD, encoding a solid as an interlinked graph of vertices, edges, loops, and faces—each equipped with exact parametric geometry (Xu et al., 2 Dec 2025). The requirements of watertightness (topological closure) and high geometric precision are fundamental for downstream tasks such as finite element simulation and manufacturing (CAM). Traditional generation approaches typically involve multi-stage decomposition—separate neural networks for vertices, edges, faces, followed by post-hoc "sewing"—leading to cumulative errors and fractured inductive priors (Xu et al., 2 Dec 2025, Jayaraman et al., 2022, Xu et al., 2024). Direct, fully neural B-Rep synthesis remained a central challenge due to the tight coupling between discrete combinatorial structure and freeform geometry.
2. Key Model Architecture and Tokenization
AutoBrep is structured as a single-stage, decoder-only Transformer with 16 self-attention layers, 32 attention heads, and a 2048-dimensional hidden size, operating on token sequences up to length 3 000 (Xu et al., 2 Dec 2025). Its core innovation is a unified tokenization scheme that linearizes both geometry and topology:
- Geometric primitives (faces and edges) are sampled as uniform grids ( for faces, for edges) in intrinsic parameter space. Each is compressed via a deep autoencoder and quantized by finite-scalar quantization (FSQ) into discrete tokens.
- Each face emits 4 geometry tokens; each edge emits 2.
- Axis-aligned bounding boxes are quantized into 1024 bins per axis, producing 6 coordinate tokens per primitive.
- Topological relationships are embedded by breadth-first traversal (BFT) of the face adjacency graph.
- Novel "local reference tokens" encode the topological references: when an edge connects a new face to a previously generated face within a two-level window, only a small, local vocabulary is needed to unambiguously specify adjacency.
- Special tokens—such as
<BOS>,<FACE_END>,<LEVEL>, complexity meta-tokens—demarcate sequence structure and support conditional control.
Sequence generation is natively autoregressive, with next-token prediction proceeding strictly according to BFT ordering, which captures both the hierarchical structure (faces followed by their bounding edges) and local topology of the solid.
3. Training Objectives, Regularization, and Inference
Training optimizes next-token cross-entropy loss over all tokens in the sequence, modeling (Xu et al., 2 Dec 2025). Additional mechanisms include:
- Level-based attention dropout (randomly masking tokens more than two BFT levels old) to encourage local correlation.
- Data augmentation by random rotation and stochastic dropping of complexity meta tokens.
For autocompletion, user-provided substructures are encoded as level 0 with a dedicated "unassigned" reference token (T_u). Inference proceeds by autoregressive sampling (top-, , temperature 1.0) with key-value caching for computational efficiency. Only tokens in the current local reference window are considered valid for topological referencing, reducing vocabulary size and improving memory locality.
4. Experimental Protocols and Quantitative Results
AutoBrep is trained and evaluated on the ABC-1M dataset: approximately 1.3 million unique STEP solids stratified by face count (up to 100 faces and 1 000 edges), as well as the ABC-Constraint subset for interface-driven autocompletion (Xu et al., 2 Dec 2025). Performance metrics include:
| Model | COV (%) | MMD (×10⁻²) | JSD (×10⁻²) | Valid (%) | Inference (s/face) |
|---|---|---|---|---|---|
| BrepGen | 66.8 | 1.65 | 1.12 | 51.2 | 1.0–2.3 |
| HoLa | 61.9 | 1.75 | 1.40 | 50.0 | 2.0 |
| DTGBrepGen* | 69.7 | 1.61 | 1.25 | 58.9 | 1.2 |
| AutoBrep | 71.5 | 1.45 | 0.97 | 70.8 | 0.46 |
Key empirical findings:
- AutoBrep achieves the highest validity (70.8%), surpassing baselines by over 20% in solids >50 faces and halving inference time.
- Geometry quantization via FSQ substantially outperforms VQ-VAE (per-point RMSE for faces: 0.075 vs. 0.338).
- Validity remains ~50% even at the largest tested solid sizes (100 faces), with consistent superiority in coverage (COV) and distributional fidelity (MMD/JSD).
- Ablation experiments reveal the necessity of local reference windows and BFT ordering: removing either degrades validity and geometric coverage significantly.
Qualitatively, AutoBrep generates solids exhibiting realistic manufacturing details (e.g., fillets, chamfers, complex cutouts) with a high degree of watertightness even at high complexity (Xu et al., 2 Dec 2025).
5. Extensions: Autocompletion, Editing, and User Control
AutoBrep's unified tokenization enables seamless support for partial B-Rep autocompletion and user-controllable generation. Given user-specified interface faces or substructures:
- The model fills in missing geometry and topology to yield watertight completions, preserving exact user input.
- Complexity meta-tokens bias toward simple or highly detailed output regimes.
- The local topology window can be restricted to enforce specified connectivity patterns.
Proposed application scenarios include CAD-aware interactive UIs, supporting commands such as "fill this hole" or "expand this flange" by modifying a small subset of tokens (Xu et al., 2 Dec 2025).
6. Comparison to Related AutoBRep Paradigms
AutoBrep advances over previous direct B-Rep models such as DTGBrepGen, BrepGen, and SolidGen by (a) integrating geometry and topology into a single token stream, (b) leveraging a purely autoregressive sequence model rather than a multi-stage pipeline, and (c) enabling scalable inference for solids at industrial complexity without cascading errors or validity collapse (Xu et al., 2 Dec 2025, Li et al., 17 Mar 2025, Xu et al., 2024, Jayaraman et al., 2022). Recent alternatives such as BrepGPT adopt related single-stage, holistic tokenization but differ in locality representation and quantization (Voronoi Half-Patch and dual VQ-VAEs) (Li et al., 27 Nov 2025); BrepARG proposes a holistic token sequence using hierarchical block construction (Li et al., 23 Jan 2026). Masked autoencoder paradigms (BRepMAE, (Yao et al., 26 Feb 2026), and hierarchical graph transformers (Li et al., 16 Mar 2026)) focus on self-supervised representation learning rather than full generative synthesis.
7. Limitations and Future Directions
Remaining challenges for AutoBrep include accurate modeling of ultra-thin features (threads, shells), where uniform point grids and coarse quantization can degrade precision. Prospective directions involve adaptive sampling strategies, hybrid continuous–discrete representations, and integrating constraints for strict manifold validity. Embedding these models in interactive CAD environments and extending to multi-modal conditional generation (sketch → B-Rep, text → B-Rep, point cloud → B-Rep) are active areas of research (Xu et al., 2 Dec 2025).
AutoBrep establishes a new paradigm for end-to-end neural CAD generation, demonstrating that with a unified token vocabulary, a single Transformer can synthesize high-fidelity, watertight B-Reps robustly and efficiently, with powerful modes for autocompletion and user-guided editing (Xu et al., 2 Dec 2025).