Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
175 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Correcting Biased Centered Kernel Alignment Measures in Biological and Artificial Neural Networks (2405.01012v1)

Published 2 May 2024 in q-bio.NC and cs.CV

Abstract: Centred Kernel Alignment (CKA) has recently emerged as a popular metric to compare activations from biological and artificial neural networks (ANNs) in order to quantify the alignment between internal representations derived from stimuli sets (e.g. images, text, video) that are presented to both systems. In this paper we highlight issues that the community should take into account if using CKA as an alignment metric with neural data. Neural data are in the low-data high-dimensionality domain, which is one of the cases where (biased) CKA results in high similarity scores even for pairs of random matrices. Using fMRI and MEG data from the THINGS project, we show that if biased CKA is applied to representations of different sizes in the low-data high-dimensionality domain, they are not directly comparable due to biased CKA's sensitivity to differing feature-sample ratios and not stimuli-driven responses. This situation can arise both when comparing a pre-selected area of interest (e.g. ROI) to multiple ANN layers, as well as when determining to which ANN layer multiple regions of interest (ROIs) / sensor groups of different dimensionality are most similar. We show that biased CKA can be artificially driven to its maximum value when using independent random data of different sample-feature ratios. We further show that shuffling sample-feature pairs of real neural data does not drastically alter biased CKA similarity in comparison to unshuffled data, indicating an undesirable lack of sensitivity to stimuli-driven neural responses. Positive alignment of true stimuli-driven responses is only achieved by using debiased CKA. Lastly, we report findings that suggest biased CKA is sensitive to the inherent structure of neural data, only differing from shuffled data when debiased CKA detects stimuli-driven alignment.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell., 35(8):1798–1828, aug 2013. doi: 10.1109/TPAMI.2013.50.
  2. img2fmri: a python package for predicting group-level fmri responses to visual stimuli using deep neural networks. Aperture Neuro, 3, October 2023. doi: 10.52294/001c.87545.
  3. Algorithms for learning kernels based on centered alignment. J. Mach. Learn. Res., 13(1):795–828, mar 2012.
  4. Aligning model and macaque inferior temporal cortex representations improves model-to-human behavioral alignment and adversarial robustness. In The Eleventh International Conference on Learning Representations, 2023.
  5. Deceiving the CKA similarity measure in deep learning. In NeurIPS ML Safety Workshop, 2022a.
  6. On the inadequacy of CKA as a measure of similarity in deep learning. In ICLR 2022 Workshop on Geometrical and Topological Representation Learning, 2022b.
  7. Reliability of CKA as a similarity measure in deep learning. In The Eleventh International Conference on Learning Representations, 2023.
  8. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp.  248–255. Ieee, 2009.
  9. Improved object recognition using neural networks trained to mimic the brain’s statistical properties. Neural Networks, 131:103–114, 2020. doi: https://doi.org/10.1016/j.neunet.2020.07.013.
  10. System identification of neural systems: If we got it right, would we know? (arXiv:2302.06677), February 2023. arXiv:2302.06677 [cs, q-bio].
  11. Deep residual learning for image recognition. CoRR, abs/1512.03385, 2015.
  12. Things-data, a multimodal collection of large-scale datasets for investigating object representations in human brain and behavior. eLife, 12:e82580, February 2023. doi: 10.7554/eLife.82580.
  13. Similarity of neural network models: A survey of functional and representational measures, 2023.
  14. Atherosclerosis and liver inflammation induced by increased dietary cholesterol intake: a combined transcriptomics and metabolomics analysis. Genome Biology, 8(9):R200, 2007. ISSN 14656906. doi: 10.1186/gb-2007-8-9-r200.
  15. Similarity of neural network representations revisited. In Proceedings of the 36th International Conference on Machine Learning, volume 97, pp.  3519–3529. PMLR, 09–15 Jun 2019.
  16. Representational similarity analysis - connecting the branches of systems neuroscience. Frontiers in Systems Neuroscience, 2, 2008.
  17. Cornet: Modeling the neural mechanisms of core object recognition. bioRxiv, 09/2018 2018.
  18. A geometric understanding of deep learning. Engineering, 6(3):361–374, 2020. doi: https://doi.org/10.1016/j.eng.2019.09.010.
  19. Individual differences among deep neural network models. Nature Communications, 11(1):5725, November 2020. doi: 10.1038/s41467-020-19632-w.
  20. Thingsvision: A python toolbox for streamlining the extraction of activations from deep neural networks. Frontiers in Neuroinformatics, 15:45, 2021. doi: 10.3389/fninf.2021.679838.
  21. Do wide and deep networks learn the same things? uncovering how neural network representations vary with width and depth. In International Conference on Learning Representations, 2021.
  22. Improving the accuracy and robustness of cnns using a deep cca neural data regularizer, 2022.
  23. Improving the accuracy of single-trial fmri response estimates using glmsingle. eLife, 11:e77599, November 2022. doi: 10.7554/eLife.77599.
  24. P. Robert and Y. Escoufier. A unifying tool for linear multivariate statistical methods: The rv-coefficient. Journal of the Royal Statistical Society Series C: Applied Statistics, 25(3):257–265, December 2018. doi: 10.2307/2347233.
  25. Brain-score: Which artificial neural network for object recognition is most brain-like? bioRxiv preprint, 2018.
  26. Integrative benchmarking to advance neurally mechanistic models of human intelligence. Neuron, 2020.
  27. Matrix correlations for high-dimensional data: the modified rv-coefficient. Bioinformatics, 25(3):401–405, December 2008. doi: 10.1093/bioinformatics/btn634.
  28. Deep learning in neuroimaging: Overcoming challenges with emerging approaches. Frontiers in Psychiatry, 13, 2022. doi: 10.3389/fpsyt.2022.912600.
  29. Supervised feature selection via dependence estimation. In Proceedings of the 24th International Conference on Machine Learning, ICML ’07, pp.  823–830, New York, NY, USA, 2007. Association for Computing Machinery. ISBN 9781595937933. doi: 10.1145/1273496.1273600.
  30. Getting aligned on representational alignment, 2023.
  31. Partial distance correlation with methods for dissimilarities. The Annals of Statistics, 42(6):2382–2412, 2014. doi: 10.1214/14-AOS1255.
  32. The effect of task and training on intermediate representations in convolutional neural networks revealed with modified rv similarity analysis. In 2019 Conference on Cognitive Computational Neuroscience, 2019. doi: 10.32470/CCN.2019.1300-0. arXiv:1912.02260 [cs, stat].
  33. Where is human v4? predicting the location of hv4 and vo1 from cortical folding. Cerebral Cortex, 24(9):2401–2408, April 2013.
  34. Daniel L K Yamins and James J DiCarlo. Using goal-driven deep learning models to understand sensory cortex. Nature Neuroscience, 19(3):356–365, March 2016.
  35. Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proceedings of the National Academy of Sciences, 111(23):8619–8624, 2014.
  36. A general index for linear and nonlinear correlations for high dimensional genomic data. BMC Genomics, 21(1):846, November 2020. doi: 10.1186/s12864-020-07246-x.

Summary

  • The paper demonstrates that biased CKA produces misleading high similarity scores in low-data, high-dimensional scenarios.
  • It employs comparative experiments with fMRI, MEG, and CNN layer data to validate the robustness of debiased CKA.
  • The findings imply that debiased CKA enhances both neuroimaging analysis and the alignment accuracy of neural network models.

A Deeper Dive into the Pitfalls of Biased CKA in Neural Data Analysis

Overview of the Research

Recent research has critically evaluated the efficacy of Centered Kernel Alignment (CKA) as an alignment metric for comparing activations from biological and artificial neural networks. While biased CKA has been a popular choice, this paper exposes its shortcomings, especially in cases involving low-data and high-dimensionality scenarios, typical of neural data like fMRI or MEG. The findings throw light on the advantages of using debiased CKA to achieve more accurate and stimuli-driven responses. Here's a walkthrough of what was discovered, the experimental approach, and the implications of these findings.

The Trouble with Biased CKA

The heart of the problem with biased CKA lies in its sensitivity to the ratio of features to samples. This intrinsic characteristic can lead to misleadingly high similarity scores for random matrices, which do not genuinely reflect proper alignment between biological responses and model architectures:

  • Erroneous High Similarity: Biased CKA can generate high similarity scores even between completely random matrices, especially if these matrices have a large number of columns (features) relative to rows (samples).
  • False Positive Alignments: The biased version of CKA may suggest an alignment between different layers of a neural network and regions of interest (ROIs) in the brain, which could be falsely attributed to similar processing of stimuli.

Debiasing the CKA

To combat the vulnerabilities of biased CKA, researchers have refined the metric to a debiased version, which demonstrates resilience to the pitfalls of its predecessor:

  • Robust to Dimensionality Issues: The debiased CKA doesn't produce artificially high similarities when faced with varying dimensions across the datasets.
  • Genuine Stimuli-driven Responses: Only the debiased CKA has shown consistency in representing actual, stimuli-driven neural responses, making it a more reliable tool in neuroimaging studies.

Experimental Insights

To bolster their arguments, the researchers conducted comprehensive experiments across different conditions using fMRI and MEG data; along with ANN layers from popular CNN models like ResNet18 and CORnet-S:

  1. Sensitivity Tests: Trials where both random and structured (real) data were exposed to biased and debiased CKA revealed significant flaws in the biased approach. For instance, shuffling sample-feature pairs did not significantly alter biased CKA results, underscoring its insensitivity to the actual stimuli-driven changes.
  2. Comparison Across Networks: Analyses of neural data in relation to layers of CNNs showed that debiased CKA is sensitive to true variations in stimuli, which biased CKA often misconstrued as random similarities.
  3. Real versus Random: Testing with a mix of real and randomized data further confirmed that debiased CKA could discern genuine similarities driven by real stimuli, which biased CKA could not distinguish from random data alignments.

Theoretical and Practical Implications

The paper sheds light not only on the theoretical understanding of neural alignment metrics but also poses significant practical implications for designing more effective neural network models and analyzing neuroimaging data:

  • Enhanced Model Accuracy: Using debiased CKA could help in developing neural network models that are better aligned with human brain responses, which may enhance the accuracy and robustness against adversarial attacks.
  • Better Neuroimaging Analysis: This metric can significantly improve the reliability of neuroimaging data analyses, directly impacting clinical and research outcomes in neuroscience.

Looking Ahead: Future Developments in AI

Reflecting on these findings, one can anticipate further refinements in neural alignment metrics. The robust attributes of debiased CKA could pave the way for new methodologies that not only tackle the dimensional disparities in data but also enhance the fidelity of cross-domain neural data analysis. Moreover, this could be a stepping stone towards creating AI systems that genuinely mimic human brain processes, a key pursuit in cognitive neuroscience and artificial intelligence.