Symmetry Breaking and Equivariant Neural Networks (2312.09016v2)
Abstract: Using symmetry as an inductive bias in deep learning has been proven to be a principled approach for sample-efficient model design. However, the relationship between symmetry and the imperative for equivariance in neural networks is not always obvious. Here, we analyze a key limitation that arises in equivariant functions: their incapacity to break symmetry at the level of individual data samples. In response, we introduce a novel notion of 'relaxed equivariance' that circumvents this limitation. We further demonstrate how to incorporate this relaxation into equivariant multilayer perceptrons (E-MLPs), offering an alternative to the noise-injection method. The relevance of symmetry breaking is then discussed in various application domains: physics, graph representation learning, combinatorial optimization and equivariant decoding.
- Machine learning for combinatorial optimization: a methodological tour d’horizon. European Journal of Operational Research, 290(2):405–421, 2021.
- Geometric deep learning: Grids, groups, graphs, geodesics, and gauges. arXiv preprint arXiv:2104.13478, 2021.
- Alan F Chalmers. Curie’s principle. The British Journal for the Philosophy of Science, 21(2):133–148, 1970.
- Pierre Curie. Sur la symétrie dans les phénomènes physiques, symétrie d’un champ électrique et d’un champ magnétique. Journal de physique théorique et appliquée, 3(1):393–415, 1894.
- A practical method for constructing equivariant multilayer perceptrons for arbitrary matrix groups. In International conference on machine learning, pages 3318–3328. PMLR, 2021.
- The symmetry perspective: from equilibrium to chaos in phase space and physical space, volume 200. Springer Science & Business Media, 2002.
- The elements of statistical learning: data mining, inference, and prediction, volume 2. Springer, 2009.
- Equivariant networks for crystal structures. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=0Dh8dz4snu.
- Equivariance with learned canonicalization functions. In International Conference on Machine Learning, pages 15546–15566. PMLR, 2023.
- Expressive sign equivariant networks for spectral geometric learning. In ICLR 2023 Workshop on Physics for Machine Learning, 2023.
- Graph normalizing flows. Advances in Neural Information Processing Systems, 32, 2019.
- Object-centric learning with slot attention. Advances in Neural Information Processing Systems, 33:11525–11538, 2020.
- Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th international conference on machine learning (ICML-10), pages 807–814, 2010.
- Charles C Pinter. A book of abstract algebra. Courier Corporation, 2010.
- Equivariance through parameter-sharing. In International Conference on Machine Learning, pages 2892–2901. PMLR, 2017.
- E (n) equivariant graph neural networks. In International conference on machine learning, pages 9323–9332. PMLR, 2021.
- Your dataset is a multiset and you should compress it like one. In NeurIPS 2021 Workshop on Deep Generative Models and Downstream Applications, 2021.
- J. Shawe-Taylor. Building symmetries into feedforward networks. In 1989 First IEE International Conference on Artificial Neural Networks, (Conf. Publ. No. 313), pages 158–162, 1989.
- Finding symmetry breaking order parameters with euclidean neural networks. Phys. Rev. Research, 3:L012002, Jan 2021. 10.1103/PhysRevResearch.3.L012002. URL https://link.aps.org/doi/10.1103/PhysRevResearch.3.L012002.
- Top-n: Equivariant set and graph generation without exchangeability. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=-Gk_IPJWvk.
- Hermann Weyl. Symmetry. In Symmetry. Princeton University Press, 1952.
- Multiset-equivariant set prediction with approximate implicit differentiation. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=5K7RRqZEjoS.