Deep Learning-based Measurement Matrix for Phase Retrieval
- The paper introduces a learnable Fourier measurement operator within an unrolled deep architecture that jointly optimizes sensing and recovery for phase retrieval.
- It employs iterative subgradient descent and learned proximal mappings to enforce data fidelity and sparsity, thereby improving reconstruction accuracy.
- Experiments show that DLMMPR consistently outperforms baseline methods in PSNR and SSIM, especially under high noise conditions.
The Deep Learning-based Measurement Matrix for Phase Retrieval (DLMMPR) framework fundamentally redefines phase retrieval by jointly optimizing the measurement operator and recovery modules in an end-to-end deep learning architecture. By parameterizing the multiplexed Fourier measurement process as learnable tensors and coupling this with unrolled subgradient descent and learned proximal mappings, DLMMPR achieves robust signal recovery under diverse noise conditions and outperforms prevailing state-of-the-art methodologies such as DeepMMSE and PrComplex in both quantitative and qualitative dimensions (Liu et al., 16 Nov 2025).
1. Measurement Matrix Parameterization
DLMMPR departs from traditional fixed coded-diffraction pattern (CDP) designs, which utilize a stack of masked Fourier transforms, by embedding the entire multi-mask Fourier measurement operator as a learnable parameter within the deep network. The conventional CDP operator is given by
where is the discrete Fourier matrix, are diagonal matrices with entries sampled on the complex unit circle, and denotes noise. DLMMPR replaces the fixed with a trainable matrix , such that at the -th stage:
Both and its adjoint are treated as weight tensors, initialized to , and updated via backpropagation with all other network parameters. This learnable parameterization facilitates joint optimization of the sensing process and iterative recovery procedure.
2. Unrolled Recovery Structure
DLMMPR utilizes K=7 unrolled ISTA-like iterations, each comprising a learnable subgradient descent (SGD) module and a learned proximal-projection mapping (PPM). At each stage :
- Subgradient-Descent (SGD) Module: Enforces data fidelity using the non-differentiable cost
with subgradient
and update
- Proximal-Projection Mapping (PPM) Module: Enforces sparsity in a learned transform domain , parameterized by a CNN block and its left-inverse .
Soft-thresholding is applied to sparse codes:
Thus, each stage combines measurement consistency (SGD) and learned sparsification (PPM) in a hybrid iterative framework.
3. End-to-End Training and Optimization
Supervised training employs batches of image patches and simulated measurements , where are the initial CDP masks. The objective is to minimize the mean squared-error:
Gradients with respect to all parameters are computed via backpropagation and updated using the Adam optimizer. DLMMPR simultaneously fits recovered signals to ground truth, enforcing both measurement-model consistency and accurate reconstruction through learned sparsity in transform domains.
4. Network Architecture and Training Protocol
- Unrolled Stages: Seven (K=7), no weight sharing across stages.
- PPM Modules: Each stage’s transform comprises:
- convolution,
- Convolutional Block Attention Module (CBAM),
- ConvU (three convolution+ReLU layers),
- Fourier Residual Block (ResFBlock).
The inverse chain mirrors these sub-modules in reverse, wrapped in a residual skip connection.
Initialization: and from Fourier operator , Xavier initialization for CNN weights, thresholds , step sizes .
- Training: PyTorch-based, Adam optimizer (, ), initial , batch size 10, 150 epochs. Input patches are , normalized to .
- Datasets: 6,000 training patches from BSD400 and PASCAL-VOC. Testing on prDeep “Set12”: UNT-6 and NT-6. Poisson noise levels .
5. Quantitative Evaluation and Empirical Performance
DLMMPR's performance is benchmarked against DeepMMSE and PrComplex (Liu et al., 16 Nov 2025). The quantitative results for Poisson noise levels on UNT-6 and NT-6 test datasets are summarized:
| Noise () | DeepMMSE | PrComplex | DLMMPR |
|---|---|---|---|
| 9 | 40.26/0.98, 39.45/0.98 | 41.11/0.98, 39.78/0.99 | 41.89/0.98, 40.41/0.99 |
| 27 | 33.03/0.94, 32.32/0.93 | 34.64/0.93, 33.49/0.94 | 35.23/0.95, 33.51/0.95 |
| 81 | 26.64/0.81, 25.41/0.78 | 28.27/0.84, 26.55/0.82 | 29.70/0.86, 27.44/0.83 |
Across all regimes, DLMMPR achieves the highest PSNR and SSIM, with typical gains exceeding +1 dB, especially at high noise levels. Visual assessment affirms preservation of sharper edges and finer textures, and superior suppression of speckle noise compared to competing methods.
6. Theoretical Insights, Convergence, and Limitations
DLMMPR is informed by LISTA/ISTA convergence theory and the weight-coupling principle, also leveraged in prior work on unfolded algorithms for phase retrieval (Naimipour et al., 2020). Although explicit proofs are absent, the architecture leverages theoretical rationale from ISTA-Net (Theorem 1) to justify the -norm approximation in the proximal step. Empirical evidence shows rapid convergence (network saturation at 150 epochs and 150 unrolled iterations), with stable performance across noise regimes.
DLMMPR's computational overhead scales with image size due to its multi-stage network structure; current instantiation assumes masks. Anticipated future work includes lightweight models and distributed optimization for large-scale imaging, and generalization of learned matrix paradigms to broader inverse problems.
7. Relationship to Prior Unfolded and Deep Measurement Matrix Approaches
Prior frameworks, such as Unfolded Phase Retrieval (UPR) (Naimipour et al., 2020), employ deep unfolding to jointly learn sensing matrices and solver parameters via interpretable iterative networks. DLMMPR advances this by generalizing the measurement operator to parameterized Fourier-like transforms and embedding learned proximal mappings via CNN modules. Both approaches advocate joint, end-to-end training that enhances measurement efficiency, convergence rate, and recovery accuracy, substantiating the benefit of data-driven measurement matrix optimization in phase retrieval.
A plausible implication is that learned measurement operators, when coupled with unrolled and interpretable recovery networks, can systematically improve imaging performance even beyond phase retrieval, motivating further methodological exploration in computational imaging and signal processing.