Normalized Bures Similarity (NBS) Overview
- Normalized Bures Similarity (NBS) is a metric that quantifies neural representation similarity using quantum-information fidelity and kernel summary statistics.
- It achieves key invariances—including orthogonal rotation, translation, permutation, and scale—by aligning geometric (Riemannian) and statistical (quantum-information) perspectives.
- Efficient computational methods such as nuclear norm-based procedures and differentiable optimization make NBS practical for comparing both artificial and biological neural systems.
Normalized Bures Similarity (NBS) is a geometric similarity measure for comparing neural representations, based on the quantum-information fidelity between covariance matrices of activation patterns. Unlike methods based on explicit unit-wise mapping, NBS quantifies similarity in kernel summary statistics, conferring invariance to orthogonal rotations, permutations, translational shifts, and scale. Its foundational identity reveals deep connections to both Riemannian geometry and quantum-information theory, allowing unification of mapping-based and kernel-based similarity frameworks. NBS enjoys properties including metric validity, mapping-free computation, and scale-rotation invariance, and further admits efficient computational schemes and differentiable optimization.
1. Mathematical Definition
Given neural activations from two systems over stimuli, stored in matrices and %%%%2%%%%, the centered linear kernel (stimulus-by-stimulus covariance) matrices are
where is the centering matrix. The fidelity between two positive semidefinite (PSD) matrices is
The normalized Bures similarity is then defined by
Alternatively,
where denotes the nuclear norm (sum of singular values). This construction is equivalent to the cosine of the Riemannian shape distance between centered neural configurations: The equivalence between NBS as kernel fidelity and as shape alignment is established rigorously (see (Harvey et al., 2023)).
2. Geometric and Statistical Interpretations
NBS is invariant under permutations and orthogonal rotations of neuron axes; centering removes translational degrees of freedom, and normalization ensures scale invariance. From the geometric perspective, NBS measures the cosine of the Riemannian (geodesic) angle between two centered configurations, corresponding to optimal Procrustes alignment in the shape manifold. Statistically, and are empirical covariance matrices and measures quantum-information fidelity. The associated Bures distance
is the 2-Wasserstein distance between zero-mean Gaussians with covariance and . By Uhlmann’s theorem,
so NBS represents the maximum normalized Hilbert-Schmidt overlap achievable for neural activations with fixed covariances (Harvey et al., 2023).
3. Computational Procedures
NBS can be computed in two equivalent ways:
- Kernel Fidelity Method:
- Center data: , .
- Compute , .
- Eigendecompose to get .
- Form and compute .
- Calculate and denominator .
- Output NBS.
- Cross-Covariance/Nuclear Norm Method:
- Center with .
- Compute , , .
- Obtain singular values of , set .
- Compute traces , .
- Output NBS.
For , the nuclear norm method is computationally preferable. As code, NBS amounts to mean-centering, computing cross-covariance, applying SVD, and normalizing by the geometric mean of marginal nuclear norms (Cloos et al., 9 Jul 2024):
1 2 3 4 5 6 7 8 9 10 |
def NBS(X, Y, epsilon=1e-6): Xc = X - X.mean(axis=0) Yc = Y - Y.mean(axis=0) M = Xc.T @ Yc num = np.sum(np.linalg.svd(M, compute_uv=False)) Cx = Xc @ Xc.T Cy = Yc @ Yc.T denom = np.sqrt(np.sum(np.linalg.svd(Cx, compute_uv=False)) * np.sum(np.linalg.svd(Cy, compute_uv=False))) + epsilon return num / denom |
4. Sensitivity to Principal Components
NBS exhibits linear sensitivity to principal component (PC) variances. If and a copy is constructed by scrambling the left singular vector (preserving variance ), then
where are the eigenvalues of (Cloos et al., 9 Jul 2024). Thus the decrease in NBS upon destroying PC is linear in its variance. By contrast, CKA’s dependence is quadratic (), making NBS more sensitive to mid-range PCs than CKA, but less so than angular Procrustes, which is most strongly sensitive to low-variance directions.
5. Metric Properties, Theorems, and Comparison to Other Measures
NBS provides a true metric (satisfying symmetry and triangle inequality) over stimulus–response geometries: The associated Bures distance is dual to the Procrustes “size-and-shape” distance : valid for neural configurations of unequal width (Harvey et al., 2023). Asymptotically, with or , the normalized Bures distance converges to its limiting form under the law of large numbers. Empirically, NBS values between random and real data start near 0, approach 0.98–1.0 under optimization, and dataset-dependent thresholds exist for meaningful “task-relevant” encoding.
Compared to other metrics:
- RSA compares vectorized representational dissimilarity matrices, does not define a true metric, and ignores global scaling.
- CKA computes
and is the cosine of the Hilbert–Schmidt angle. Unlike NBS, CKA neither exploits PSD geometry nor satisfies the triangle inequality, and it can be insensitive to alignment of dominant covariance subspaces. Tight bounds relate CKA and NBS:
and empirical discrepancies can be two- to three-fold.
- CCA fits optimal linear mappings to maximize correlation in a shared subspace; mapping-based, affine-invariant, requiring generalized eigenproblems, and returns multiple coefficients. NBS is mapping-free, giving a single overlap scalar.
NBS should be preferred when metric validity, PSD-manifold geometry, and mapping-free similarity are desired.
6. Empirical Behavior, Optimization, and Interpretation
Differentiable optimization allows maximizing NBS directly. Under optimization, synthetic data initially learns the highest-variance PC of the target dataset, with NBS capturing lower-variance PCs more rapidly than CKA but less so than angular Procrustes. While high NBS scores () can be achieved, they do not guarantee encoding of all task-relevant dimensions; “overfitting” to highest-variance PCs is possible. No single threshold for “good” NBS exists; the appropriate value depends on dataset structure (e.g., $0.6$ NBS suffices for high task-decoding accuracy in some prefrontal recordings, while others require ).
Empirical scatter plots show that CKA and NBS are correlated, but substantial envelope width remains due to kernel ranks and matrix square-root non-commutativity. Joint optimization experiments indicate that high angular Procrustes score entails high NBS and CKA, but not vice-versa; very high CKA or NBS can still omit lower-variance, task-relevant structure (Cloos et al., 9 Jul 2024).
7. Applications and Limitations
NBS is applicable for quantifying similarity between neural representations, including artificial and biological systems, without explicit neuron correspondences. It is effective for characterizing neural encoding overlap under PSD-manifold geometry. NBS’s mapping-free nature, scale and rotation invariance, and metric validity offer distinct advantages over ad hoc measures. However, NBS (like kernel-based measures) is susceptible to “overfitting” dominant PCs and may not ensure task-relevant dimensionality recovery. For practitioners, careful interpretation is required, especially when optimizing representations: additional analysis concerning encoding of low-variance dimensions and task variables is necessary for robust assessment.
NBS unifies geometric (shape manifold) and statistical (quantum-information and Wasserstein) perspectives on neural similarity, and its properties are increasingly prominent in comparative studies of neural representations in deep learning and neuroscience (Harvey et al., 2023, Cloos et al., 9 Jul 2024).