CPCR: Cross-Pyramid Consistency for Segmentation
- CPCR is a semi-supervised framework that improves medical image segmentation by enforcing cross-pyramid consistency across dual decoders.
- It employs a dual-branch pyramid network with distinct upsampling methods and scale-specific perturbations to generate multi-scale auxiliary predictions.
- CPCR integrates multiple loss terms including Dice loss, KL-divergence consistency, and entropy minimization, yielding competitive DSC, IoU, and enhanced boundary accuracy.
Cross-Pyramid Consistency Regularization (CPCR) is a semi-supervised learning framework designed to improve medical image segmentation accuracy by leveraging unlabeled data through a hybrid consistency-based approach applied to a dual-branch pyramid network. CPCR operates by enforcing cross-decoder and cross-scale knowledge distillation between two slightly different decoders, each producing multiscale auxiliary predictions with targeted feature perturbations, and integrates established consistency and uncertainty minimization objectives. This strategy effectively regularizes deep hierarchical features and demonstrates competitive performance with recent state-of-the-art methods while using a compact model architecture.
1. Dual Branch Pyramid Network (DBPNet) Architecture
DBPNet serves as the backbone for CPCR, implementing a four-level U-Net-style encoder and two structurally distinct decoders. The encoder consists of four sequential blocks of two 3×3 convolutions with ReLU activations, followed by 2×2 max-pooling steps, yielding progressively downsampled feature maps.
Each decoder (TR-branch and UP-branch) mirrors the encoder's structure but differs in its upsampling mechanism:
- TR-branch: Employs transpose-convolution with learned kernels.
- UP-branch: Utilizes 1×1 convolution followed by bilinear upsampling.
Skip connections link encoder blocks to their corresponding decoder levels. At the first three decoder levels (scales ), each branch produces an auxiliary prediction through a 3×3 convolution, upsampling to input resolution, and softmax activation. A scale-specific perturbation module (dropout, feature-dropout, or Gaussian noise) is applied to features before the auxiliary head, with perturbations differing across both scale and branch. The fourth decoder level outputs the main prediction without auxiliary heads. Thus, each branch provides a pyramid of probability maps for .
2. Cross-Pyramid Consistency Regularization: Mathematical Foundations
The CPCR regularization term acts on auxiliary predictions at scales . For each, soft-labeling is performed using temperature-scaled softmax:
where is the logits tensor, denotes the temperature (set to 10), and indexes pixels and classes.
CPCR enforces reciprocal consistency between the two branches via symmetric KL-divergence at each auxiliary scale:
The overall CPCR loss averages these terms over three scales:
This multi-scale, cross-branch distillation framework facilitates the alignment of deep feature distributions hierarchically, supporting effective exploitation of unlabeled data and enhancing regularization.
3. Composite Loss Function and Optimization Protocol
DBPNet with CPCR employs a total loss function aggregating the following components:
- Supervised Dice Loss (): Applied to main outputs () on labeled data, summing Dice losses for both branches:
where is the ground-truth mask.
- Main Output Consistency (): Symmetric KL-divergence between main outputs of the two decoders:
- Average-Prediction Entropy Minimization (): Reduces uncertainty by minimizing pixel-wise entropy of the averaged main output prediction:
where is the number of pixels.
- CPCR Regularization (): As defined above.
The total loss is a weighted sum:
With , and as a sigmoid-shaped warm-up weight increasing from 0 to over epochs:
4. Training and Inference Workflow
Training DBPNet+CPCR proceeds as follows, using a batch size equally split between labeled and unlabeled examples. Random augmentations (flips, rotations) are applied to inputs prior to encoding. The algorithm cycles through the following steps for iterations:
- Forward pass through encoder and both decoders;
- Extraction of logits and computation of probability maps;
- Softmax (standard and temperature) transformations;
- Calculation of loss terms on appropriate data (labeled or both);
- Weighted aggregation of loss components and update of per iteration;
- Backpropagation and parameter updates via SGD (learning rate , momentum $0.9$, weight decay ).
At inference, only the UP-branch main output () is used.
Training Pseudocode
1 2 3 4 5 6 7 8 9 10 11 12 13 |
for iteration t = 1 to I: Sample B/2 labeled {(x_i, y_i)}, B/2 unlabeled {x_j} Apply random ± flip, ± rotation Encoder → features → Decoders TR, UP For d ∈ {TR, UP}, l = 1..4: Obtain logits Z_d^{(l)} Compute P_d^{(l)} = softmax(Z_d^{(l)}) Compute temperature softmax \tilde{P}_d^{(l)} Compute ℒ_sup (labeled) Compute ℒ_con^main, ℒ_um, ℒ_CPCR (all data) Calculate ℒ_total using α and λ(t) Backpropagate gradient; update θ via SGD Every 150 iterations: update λ(t_epoch) |
5. Empirical Evaluation and Comparative Analysis
Evaluation on the ACDC dataset using only 10% annotated labels demonstrates the efficacy of CPCR. The method achieves the following results:
| Model | DSC (%) | IoU (%) | 95HD (mm) | ASD (mm) |
|---|---|---|---|---|
| Baseline U-Net | 77.3 | — | — | — |
| UA-MT (2019) | 81.6 | — | — | — |
| URPC (2021) | 81.8 | — | — | — |
| MC-Net+ (2022) | 87.1 | — | — | — |
| DVCPS (2025) | 88.8 | — | — | — |
| DBPNet + CPCR | 88.11 | 79.45 | 4.12 | 1.11 |
With only two decoder branches, CPCR achieves competitive Dice Similarity Coefficient (DSC) and Intersection over Union (IoU) compared to recent works, while yielding superior boundary accuracy reflected in lower 95% Hausdorff distance and Average Surface Distance (ASD). This suggests boundary precision benefits from hierarchical cross-decoder regularization.
6. Significance and Methodological Context
CPCR extends existing paradigms in semi-supervised segmentation by infusing multi-scale reciprocal distillation into dual decoder architectures. Distinct decoding mechanisms and targeted perturbations promote diverse feature representations, while cross-pyramid KL-based regularization supports hierarchical knowledge transfer and label-efficient training. The approach combines the strengths of consistency learning and uncertainty minimization, integrating them seamlessly with supervised and self-supervised objectives through a sigmoidal weighting schedule.
A plausible implication is that the CPCR mechanism could generalize to other hierarchical tasks where multi-branch architectures and cross-scale feature alignment are advantageous. Its demonstrated competitive accuracy and compact design attest to the potential for scalable semi-supervised learning in medical imaging with limited annotations.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free