Phase-Conditioned Semantic Priors in MRI
- The paper introduces a generative model that integrates phase augmentation into MRI reconstruction to significantly improve image fidelity and robustness.
- The method employs both PixelCNN and diffusion-based priors, achieving higher PSNR and SSIM compared to magnitude-only or L1-wavelet approaches.
- Practical guidance is provided for scalable phase augmentation workflows that integrate seamlessly with linear (FISTA) and nonlinear (IRGNM) MRI reconstruction algorithms.
Phase-conditioned semantic priors are generative models specifically designed to regularize complex-valued inverse problems, notably in magnetic resonance imaging (MRI), by incorporating both magnitude and phase information within the prior learning process. Through a structured workflow—generation of phase-augmented complex training samples, training of a generative prior, and deployment as a regularizer in reconstruction—these priors enable semantic-level constraints conditioned on plausible phase, substantially improving image fidelity, robustness, and quantitative performance, especially under strong undersampling regimes (Luo et al., 2023).
1. Mathematical Foundation and Formulation
Phase-conditioned semantic priors model the empirical distribution of complex-valued MRI images. Let denote the magnitude image and the phase map. The complex-valued image is constructed as
Given undersampled multi-coil k-space data and encoding operator , the Maximum-A-Posteriori (MAP) MRI reconstruction with a generative prior is
or, equivalently,
2. Phase Augmentation Strategy
To overcome the scarcity of phase annotations in clinical datasets, phase-conditioned priors leverage phase augmentation. The workflow begins with a magnitude-only dataset and augments each sample by synthesizing corresponding phase maps . Two approaches are considered:
- Uniform phase sampling: independently.
- Diffusion-based phase sampling: sampled from a learned conditional prior , implemented via a conditional Langevin sampler with a pre-trained complex diffusion model .
For the latter, phase-augmented complex images are drawn by iteratively updating
until convergence, then extracting as the argument and as the modulus of .
3. Generative Prior Architectures and Training
Two principal generative modeling frameworks are employed:
- PixelCNN-based prior:
- Input: Two-channel images .
- Architecture: 24-layer gated causal convolutions with 64 features each, followed by a 10-component mixture of discretized logistics per pixel; M parameters.
- Loss: Negative log-likelihood,
where is the -th raster-scan pixel.
- Diffusion-based prior (Score network):
- Input: Noisy complex image at noise-level .
- Architecture: U-Net style score-based model (“Refine-Net”) with residual blocks and attention, channels [64, 128, 256]; M parameters.
- Loss: Denoising score matching (DSM),
with , .
Training is conducted on either a small complex dataset (k images) or a large phase-augmented magnitude dataset (k slices), producing corresponding priors (small) and (large).
4. Integration into MRI Reconstruction Algorithms
The learned priors are incorporated as regularizers in both linear and nonlinear MRI reconstruction schemes:
- Linear PICS (Proximal-Gradient/FISTA):
- Objective: .
- FISTA iteration:
with the proximal operator for approximated by a single gradient step.
- Nonlinear IRGNM/NLINV (Joint Image and Coil Sensitivity Estimation):
- Joint forward model solved by Gauss–Newton iterations,
with a two-stage scheme: conjugate gradient (CG) with Tikhonov prior for early iterations, then FISTA with the learned prior for refinement.
5. Quantitative Evaluation
Experimental assessments include various undersampling and reconstruction settings:
| Prior/model | PSNR (5×) | SSIM (5×) | Comments |
|---|---|---|---|
| Magnitude-only | 39 dB | 0.95 | Phase artifacts |
| Small complex | 43 dB | 0.98 | Clean magnitude and phase |
| Large complex | 44 dB | 0.99 | Fewer outliers, robust |
| -wavelet | 38 dB | 0.94 | Inferior at higher undersampling |
Further results:
- Nonlinear NLINV reconstructions reflected similar trends; magnitude-only priors introduced ghosting, especially in linear PICS.
- The large training set reduced PSNR/SSIM variance across slices, particularly at 4× and 6× undersampling.
- 3D TurboFLASH (8.2×) blind reader scores (1–5 scale): coil-combination reference 5.0, -wavelet PICS 3.2±0.6, diffusion prior PICS 4.1±0.5, NLINV 3.8±0.6.
6. Practical Guidance and Insights
- Priors trained exclusively on magnitude data are fundamentally limited in recovering phase, leading to smoothing artifacts in both magnitude and phase reconstructions.
- Phase augmentation using a small set of complex-labeled examples (employing a diffusion prior) enables the exploitation of large existing institutional databases for learning rich, complex-valued priors.
- PixelCNNs provide more accurate likelihood modeling but are computationally intensive; diffusion priors offer an effective trade-off between sample fidelity, speed of gradient evaluation (approximately 100 evaluations per reconstruction), and robustness.
- Effective protocol: phase-augment with a few thousand complex-labeled samples, generate multiple (5–10) phase realizations per magnitude slice, then train the generative prior on the expanded set.
- The learned prior can be used as a drop-in regularizer in FISTA or IRGNM frameworks through proximal-gradient integration of the learned score or log-likelihood gradient.
- Nonlinear IRGNM/NLINV is preferable at high undersampling rates (>5×) to mitigate aliasing where linear models are insufficient.
- Phase augmentation generalizes prior learning, obviating the requirement for paired magnitude-phase ground truth.
7. Significance and Implications
Phase-conditioned semantic priors expand the utility of MRI deep learning workflows by leveraging abundant magnitude-only image archives for high-fidelity, complex-valued prior construction. This facilitates improved MR image reconstruction, particularly under aggressive undersampling, without dependence on extensive paired complex datasets. A plausible implication is that this methodology could be generalized to other imaging settings where phase information is important yet underrepresented in available datasets (Luo et al., 2023).