Kernel CurvSSL: Curvature in RKHS
- The paper introduces a self-supervised framework that augments redundancy reduction by explicitly aligning local curvature in an RKHS.
- It computes discrete curvature from k-nearest neighbor graphs and aligns these metrics across augmented views using a Barlow Twins-style loss.
- The integrated loss, combining embedding and curvature objectives, leads to measurable performance gains on MNIST and CIFAR-10.
Kernel CurvSSL is a self-supervised representation learning framework that augments conventional non-contrastive redundancy-reduction objectives with explicit curvature regularization in a reproducing kernel Hilbert space (RKHS). It extends the CurvSSL family by formulating discrete local curvature with respect to a kernel-induced geometry and aligns such curvature between augmented views using a Barlow Twins-style loss on curvature-derived statistics. This approach aims to enforce not only statistical invariance and redundancy reduction, but also consistency in the local manifold geometry of learned embeddings, thereby enriching the expressiveness and utility of the resulting representation space (Ghojogh et al., 21 Nov 2025).
1. Discrete Curvature in RKHS
Kernel CurvSSL treats projected embeddings as vertices whose local geometry is quantified via discrete curvature scores. For each embedding , the nearest neighbors are determined, and their displacements relative to are given by . In the RKHS, with feature map and kernel , the local Gram matrix of the neighborhood is
Normalization to unit diagonal (mimicking projection onto the unit hypersphere) is performed:
The discrete curvature score at is defined by the sum of off-diagonal entries of :
This score can also be expressed as
and, using eigenvalues of , as
2. Curvature-based Regularizer
To regularize the local geometry, Kernel CurvSSL computes curvature scores and for two stochastic augmentations per sample in a mini-batch of size . The curvature vectors are centered and variance-normalized:
A cross-correlation matrix is formed:
whose diagonals correspond to matched samples and off-diagonals to mismatched ones. The curvature alignment loss is:
In matrix form:
This regularizer enforces view-invariance of local curvature and discourages redundancies in curvature patterns across different samples.
3. Kernel CurvSSL Loss Function
The embedding-level redundancy-reduction loss used in Kernel CurvSSL is analogous to that in Barlow Twins, applied to projected features:
with cross-correlation matrix
The embedding loss is
The total loss is a weighted sum of the embedding-level and curvature-based losses:
Explicitly,
4. Encoder–Projector Architecture and Training Workflow
Kernel CurvSSL employs a standard two-view self-supervised architecture composed of a shared encoder and a projection head . The typical training iteration proceeds as follows:
- Sample a mini-batch .
- Generate two independent stochastic augmentations for each , yielding .
- Compute representations: , and , .
- Compute per-view means and standard deviations over ; normalize to obtain .
- Form cross-correlation and evaluate .
- For each , find nearest neighbors by (Euclidean or kernel) distance among , form , normalize to , compute ; repeat for second view.
- Stack curvature scores and ; center and scale to obtain .
- Form and compute .
- Backpropagate through all differentiable operations. Neighbor selection is treated as fixed with respect to gradients; kernel evaluations allow gradients to flow from to parameters.
5. Geometric and Methodological Context
Non-contrastive self-supervised schemes such as Barlow Twins and VICReg enforce alignment of statistical moments (invariance, variance, covariance) but do not shape the local manifold geometry of embeddings. Kernel CurvSSL introduces an explicit mechanism for regularizing and aligning the bending—captured by neighborhood curvature—across augmentations. This ensures:
- Local bending consistency: If two augmentations fall nearby in the learned space, they exhibit similar local curvature patterns in their neighbor graphs.
- Collapse avoidance: By decorrelating curvature across samples, the approach mitigates degenerate solutions where the underlying manifold lacks structural diversity.
- Nonlinear geometric capture: The RKHS formulation enables modeling of complex, nonlinear local structures, enriching the representation beyond what Euclidean relationships provide.
These considerations underscore a shift from purely statistical regularization towards methods that leverage more sophisticated geometric priors.
6. Experimental Results and Comparative Performance
Kernel CurvSSL was evaluated using ResNet-18 encoders and 2-layer MLP projectors with output dimension on MNIST and CIFAR-10, using 100 and 500 pretraining epochs, respectively. Batch size, neighborhood size, and key hyperparameters were set as , , , , , with RBF kernel for kernel CurvSSL.
Standard augmentations were employed: random crop/resize + small rotation (MNIST), and random crop + flip + color jitter + grayscale (CIFAR-10). Optimization used Adam with learning rate and weight decay . For evaluation, the encoder was frozen, and a linear classifier was trained for 50 epochs (SGD).
| Method | MNIST Top-1 (%) | CIFAR-10 Top-1 (%) |
|---|---|---|
| VICReg | 95.9 | 74.5 |
| Barlow Twins | 94.9 | 73.6 |
| CurvSSL (Euclid) | 97.9 | 75.1 |
| Kernel CurvSSL | 98.4 | 76.5 |
Kernel CurvSSL outperformed both statistical regularizers and its Euclidean variant by 1–2 points (CIFAR-10) and 2–3 points (MNIST) in linear probe accuracy, indicating the efficacy of explicitly regularizing local geometry in RKHS as a complement to standard SSL approaches (Ghojogh et al., 21 Nov 2025).