Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Interpreting Equivariant Representations (2401.12588v1)

Published 23 Jan 2024 in cs.LG and stat.ML

Abstract: Latent representations are used extensively for downstream tasks, such as visualization, interpolation or feature extraction of deep learning models. Invariant and equivariant neural networks are powerful and well-established models for enforcing inductive biases. In this paper, we demonstrate that the inductive bias imposed on the by an equivariant model must also be taken into account when using latent representations. We show how not accounting for the inductive biases leads to decreased performance on downstream tasks, and vice versa, how accounting for inductive biases can be done effectively by using an invariant projection of the latent representations. We propose principles for how to choose such a projection, and show the impact of using these principles in two common examples: First, we study a permutation equivariant variational auto-encoder trained for molecule graph generation; here we show that invariant projections can be designed that incur no loss of information in the resulting invariant representation. Next, we study a rotation-equivariant representation used for image classification. Here, we illustrate how random invariant projections can be used to obtain an invariant representation with a high degree of retained information. In both cases, the analysis of invariant latent representations proves superior to their equivariant counterparts. Finally, we illustrate that the phenomena documented here for equivariant neural networks have counterparts in standard neural networks where invariance is encouraged via augmentation. Thus, while these ambiguities may be known by experienced developers of equivariant models, we make both the knowledge as well as effective tools to handle the ambiguities available to the broader community.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (45)
  1. Group invariant machine learning by fundamental domain projections. In Sophia Sanborn, Christian Shewmake, Simone Azeglio, Arianna Di Bernardo, and Nina Miolane (eds.), Proceedings of the 1st NeurIPS Workshop on Symmetry and Geometry in Neural Representations, volume 197 of Proceedings of Machine Learning Research, pp.  181–218. PMLR, December 2023.
  2. Glen E Bredon. Introduction to compact transformation groups. Academic press, 1972.
  3. Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. CoRR, abs/2104.13478, 2021. URL https://arxiv.org/abs/2104.13478.
  4. Populations of unlabelled networks: Graph space geometry and generalized geodesic principal components. Biometrika, pp.  asad024, 2023.
  5. Near-optimal signal recovery from random projections: Universal encoding strategies? IEEE transactions on information theory, 52(12):5406–5425, 2006.
  6. A program to build E(N)-equivariant steerable CNNs. In International Conference on Learning Representations, 2022a. URL https://openreview.net/forum?id=WE4qe9xlnQw.
  7. A program to build E(N)-equivariant steerable CNNs. In International Conference on Learning Representations, 2022b. URL https://openreview.net/forum?id=WE4qe9xlnQw.
  8. Group equivariant convolutional networks. In Maria Florina Balcan and Kilian Q. Weinberger (eds.), Proceedings of The 33rd International Conference on Machine Learning, volume 48 of Proceedings of Machine Learning Research, pp.  2990–2999, New York, New York, USA, 20–22 Jun 2016. PMLR. URL https://proceedings.mlr.press/v48/cohenc16.html.
  9. Li Deng. The MNIST database of handwritten digit images for machine learning research [best of the web]. IEEE signal processing magazine, 29(6):141–142, 2012.
  10. Learning meaningful representations of protein sequences. Nature communications, 13(1):1914, 2022.
  11. Statistics on stratified spaces. In Riemannian geometric statistics in medical image analysis, pp.  299–342. Elsevier, 2020.
  12. Fast graph representation learning with PyTorch Geometric. In ICLR Workshop on Representation Learning on Graphs and Manifolds, 2019.
  13. Generative adversarial networks. Communications of the ACM, 63(11):139–144, 2020.
  14. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp.  16000–16009, 2022.
  15. Multiresolution equivariant graph variational autoencoder. Machine Learning: Science and Technology, 4(1):015031, 2023.
  16. An introduction to variational autoencoders. Foundations and Trends® in Machine Learning, 12(4):307–392, 2019.
  17. Segment anything. arXiv preprint arXiv:2304.02643, 2023.
  18. Averages of unlabeled networks: geometric characterization and asymptotic behaviour. The Annals of Statistics, 48(1):514–538, 2020.
  19. MNIST handwritten digit database. http://yann.lecun.com/exdb/mnist/, 2010. URL http://yann.lecun.com/exdb/mnist/.
  20. Constrained graph variational autoencoders for molecule design. Adv. Neural Inf. Process. Syst., 31, 2018.
  21. Constrained generation of semantically valid graphs via regularizing variational autoencoders. arxiv:1809.02630, September 2018.
  22. KV Mardia and IL Dryden. The statistical analysis of shape data. Biometrika, 76(2):271–281, 1989.
  23. Invariant and equivariant graph networks. arxiv:1812.09902, December 2018a.
  24. Invariant and equivariant graph networks. arxiv:1812.09902, December 2018b.
  25. Provably powerful graph networks. arxiv:1905.11136, May 2019.
  26. Disentangling disentanglement in variational autoencoders. In Kamalika Chaudhuri and Ruslan Salakhutdinov (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp.  4402–4412. PMLR, 2019.
  27. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv:1802.03426, 2018.
  28. Manifold learning in quotient spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.  9165–9174, 2018.
  29. A graph VAE and graph transformer approach to generating molecular graphs. arxiv:2104.04345, April 2021.
  30. Permutation equivariant layers for higher order interactions. In Gustau Camps-Valls, Francisco J R Ruiz, and Isabel Valera (eds.), Proceedings of The 25th International Conference on Artificial Intelligence and Statistics, volume 151 of Proceedings of Machine Learning Research, pp.  5987–6001. PMLR, 2022.
  31. Deep k-nearest neighbors: Towards confident, interpretable and robust deep learning. arXiv:1803.04765, 2018.
  32. Automatic differentiation in pytorch. In NIPS-W, 2017.
  33. Frame averaging for invariant and equivariant network design. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=zIUyj55nXR.
  34. Quantum chemistry structures and properties of 134 kilo molecules. Sci Data, 1:140022, August 2014.
  35. Conditional constrained graph variational autoencoders for molecule design. 2020 IEEE Symposium Series on Computational Intelligence (SSCI), 2020.
  36. Enumeration of 166 billion organic small molecules in the chemical universe database GDB-17. J. Chem. Inf. Model., 52(11):2864–2875, November 2012.
  37. Manifold valued data analysis of samples of networks, with applications in corpus linguistics. The Annals of Applied Statistics, 16(1):368–390, 2022.
  38. Graphvae: Towards generation of small graphs using variational autoencoders. In Artificial Neural Networks and Machine Learning–ICANN 2018: 27th International Conference on Artificial Neural Networks, Rhodes, Greece, October 4-7, 2018, Proceedings, Part I 27, pp.  412–422. Springer, 2018.
  39. The general theory of permutation equivarant neural networks and higher order graph variational encoders. CoRR, abs/2004.03990, 2020. URL https://arxiv.org/abs/2004.03990.
  40. Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-SNE. Journal of machine learning research, 9(11), 2008.
  41. Top-N: Equivariant set and graph generation without exchangeability. arxiv:2110.02096, October 2021.
  42. General E(2)-Equivariant Steerable CNNs. In Conference on Neural Information Processing Systems (NeurIPS), 2019.
  43. Permutation-invariant variational autoencoder for graph-level representation learning. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp.  9559–9573. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/4f3d7d38d24b740c95da2b03dc3a2333-Paper.pdf.
  44. Unsupervised learning of group invariant and equivariant representations. arxiv:2202.07559, February 2022.
  45. MoleculeNet: a benchmark for molecular machine learning. Chem. Sci., 9(2):513–530, January 2018.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets