GenTract: Generative Global Brain Tractography
- GenTract is a conditional generative model for global brain tractography that uses dMRI-derived SH volumes to produce anatomically plausible streamline sets.
- It integrates diffusion-based and flow-matching paradigms to efficiently generate full tractograms while minimizing error propagation and noise sensitivity.
- The framework leverages transformer generators and VAE-based anatomical encoders to achieve state‐of-the-art precision, robust performance, and fast inference.
GenTract is a conditional generative model for global brain tractography based on diffusion magnetic resonance imaging (dMRI). It reframes the process of tractogram reconstruction as a conditional generative modeling task, learning a direct mapping from dMRI-derived spherical harmonic (SH) coefficient volumes to complete, anatomically plausible sets of streamlines. GenTract represents the first generative approach for global tractography, integrating recent advances in diffusion-based and flow-matching generative paradigms to produce full tractograms in parallel, conditioned on the global dMRI signal (Sargood et al., 17 Nov 2025).
1. Background and Problem Formulation
Tractography aims to reconstruct trajectories of white-matter fiber bundles by inferring streamlines from dMRI data. Classical local methods generate streamlines incrementally by advancing along estimated local fiber orientations. These approaches suffer from error accumulation and high false positive rates, particularly under noisy or low-resolution conditions. Global tractography methods jointly optimize entire collections of streamlines for consistency with local fiber orientations but incur significant computational costs and may converge to suboptimal solutions.
Formally, GenTract operates between spaces:
- : 3D dMRI-derived SH coefficient volumes with coefficients per voxel.
- $\mathcal{Y} = \{Y = \{s^{(i)}_{i=1}^N\},\ s^{(i)} \in \mathbb{R}^{p \times 3}\}$: tractograms comprising streamlines, each a sequence of 3D points.
The generative objective is to learn a conditional generator such that
$G(x) = \{s^{(i)}_{i=1}^N \sim p_\theta(Y|x)\},$
with the parameter vector optimized to minimize negative log-likelihood:
2. Generative Modeling Paradigms
GenTract implements two families of conditional generative models for given : diffusion-based generative modeling and flow-matching.
2.1 Diffusion-Based Generative Modeling
This approach employs a continuous-time stochastic differential equation (SDE) framework. The forward (noising) SDE is defined as
where is standard Brownian motion. Generation proceeds via the reverse-time SDE: A score network is trained to approximate using the denoising score matching objective: where and predicts .
2.2 Flow-Matching Generative Modeling
In flow-matching, a neural network learns a vector field transporting noise to data along the ODE: The training objective minimizes:
3. Model Architecture
GenTract’s architecture consists of two principal modules.
3.1 Anatomical Conditioning Encoder
For each of SH coefficient volumes, an independent variational autoencoder (VAE) is trained:
- Encoder ,
- Decoder .
Each VAE is optimized with a composite objective: The resulting latent tensors are fused via a shared 3D ResNet-style encoder and concatenated into a global context vector .
3.2 Conditional Transformer Generator
Inputs to the generator are:
- Noisy streamline ,
- Generation time ,
- Conditioning context .
Inputs are linearly projected into an embedding space with dimension and supplemented with sinusoidal position encodings and learned time embeddings. The core comprises Transformer layers with self-attention over streamline points and cross-attention to inject global anatomical context. The output is a residual 3D offset or denoised 3D point sequence.
Key architectural details include embedding dimension and usage of 10 DDIM inference steps for diffusion-based generation.
4. Training Protocol and Data
4.1 Loss Functions
The overall training objective aggregates:
- VAE per-coefficient losses: , , , ,
- Generative loss: or .
Gradients are backpropagated throughout both the transformer-based generator and the anatomical encoder.
4.2 Datasets and Augmentation
GenTract is trained on the HCP Young Adult dataset (1,042 subjects), targeting PyAFQ-filtered tractograms with 24 known bundles. The dataset is split into training (75%), validation (10%), and test (15%) partitions at the subject level.
Data augmentations include deterministic rotations of both dMRI volumes and streamlines by 15°, 30°, 45°. Synthetic corruptions are introduced at test time:
- Rician noise with on diffusion-weighted images,
- Downsampling to with added noise to simulate clinical image quality.
An external robustness evaluation was performed on TractoInferno (28 subjects) with similar degradation.
4.3 Model Implementation
- VAE initialized via fine-tuned MAISI backbone (Guo et al. 2025).
- All modules implemented in MONAI and run on NVIDIA H100 accelerators.
- Inference proceeds in batches with 10 DDIM steps.
5. Evaluation Metrics and Quantitative Results
5.1 Metrics
Given the lack of direct fiber “ground truth,” GenTract is evaluated on:
- Precision: , where TP/FP are streamlines retained/discarded by reference filtering.
- Bundle count: Number of bundles (out of 24 or 51) populated by at least one plausible streamline.
- Inference time: Mean time for full tractogram generation.
General metrics defined but not reported in the study include recall and coverage.
5.2 Comparative Performance
| BS % Precision | BS Bundles | TO-Net % Precision | |
|---|---|---|---|
| tckglobal | 0.19 | 42.91 | 17.83 |
| iFOD2 | 1.96 | 48.88 | 6.30 |
| SD Stream | 4.71 | 47.85 | 11.10 |
| DDTracking | 0.49 | 9.30 | 8.35 |
| TractOracle | 28.93 | 48.20 | 39.55 |
| GenTract | 61.95 | 36.62 | 56.35 |
- On clean HCP test data, GenTract achieves 61.95% precision versus 28.93% for TractOracle ( improvement).
- Under Rician noise, GenTract maintains 60.32% precision vs. 22.06% for TractOracle.
- On low-resolution, noisy data, GenTract’s precision is 15.73%, exceeding the next-best (TractOracle, 1.12%) by an order of magnitude.
- On external TractoInferno (low-res + noise), GenTract retains 24.94% precision (TractOracle: 9.74%).
5.3 Inference Speed
| Model Type | Mean Inference Time (s) |
|---|---|
| Classical global/local | ~2,000–3,500 |
| Deep-learning local | ~150–300 |
| GenTract | ~230 |
This suggests GenTract matches or exceeds deep local approaches in efficiency and is an order of magnitude faster than classical global methods.
6. Robustness and Qualitative Behavior
Qualitative analysis indicates that GenTract reliably retains anatomically plausible bundles across a range of acquisition conditions. For example, in the Right Superior Longitudinal Fasciculus (SLFR), GenTract maintains plausible streamline geometry under clean, noisy, and low-resolution/noisy conditions, while other methods (SD Stream, iFOD2, tckglobal) fail entirely in challenging scenarios. Error heatmaps and streamline overlays demonstrate GenTract’s robust anatomical plausibility and resistance to noise and resolution degradation.
7. Limitations and Future Directions
GenTract advances tractography by directly sampling tractograms conditioned on full-brain SH embeddings, thereby eliminating stepwise error propagation and the need for manual seeding masks. The global latent context enhances robustness against noise and resolution effects.
However, limitations include:
- Supervision is currently limited by the PyAFQ 24-bundle labeling, contributing to a modest false negative rate.
- Results depend on proxy “ground truth” and synthetic corruptions, which may introduce evaluation biases.
Future research directions include:
- Expanding the training set with more bundles or moving toward unsupervised/global training paradigms that avoid explicit streamline labels.
- Validation on real clinical datasets with genuine low-resolution and noise characteristics.
- Incorporating anatomically-informed priors during training and adopting alternative evaluation metrics such as topological fidelity and coverage.
GenTract establishes a new generative framework for global tractography, providing state-of-the-art precision on research-grade datasets and demonstrating robustness to clinically relevant data degradations (Sargood et al., 17 Nov 2025).