Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 59 tok/s
Gemini 2.5 Pro 49 tok/s Pro
GPT-5 Medium 32 tok/s Pro
GPT-5 High 33 tok/s Pro
GPT-4o 127 tok/s Pro
Kimi K2 189 tok/s Pro
GPT OSS 120B 421 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

A General Framework for Robust G-Invariance in G-Equivariant Networks (2310.18564v2)

Published 28 Oct 2023 in cs.LG, cs.AI, and cs.CV

Abstract: We introduce a general method for achieving robust group-invariance in group-equivariant convolutional neural networks ($G$-CNNs), which we call the $G$-triple-correlation ($G$-TC) layer. The approach leverages the theory of the triple-correlation on groups, which is the unique, lowest-degree polynomial invariant map that is also complete. Many commonly used invariant maps--such as the max--are incomplete: they remove both group and signal structure. A complete invariant, by contrast, removes only the variation due to the actions of the group, while preserving all information about the structure of the signal. The completeness of the triple correlation endows the $G$-TC layer with strong robustness, which can be observed in its resistance to invariance-based adversarial attacks. In addition, we observe that it yields measurable improvements in classification accuracy over standard Max $G$-Pooling in $G$-CNN architectures. We provide a general and efficient implementation of the method for any discretized group, which requires only a table defining the group's product structure. We demonstrate the benefits of this method for $G$-CNNs defined on both commutative and non-commutative groups--$SO(2)$, $O(2)$, $SO(3)$, and $O(3)$ (discretized as the cyclic $C8$, dihedral $D16$, chiral octahedral $O$ and full octahedral $O_h$ groups)--acting on $\mathbb{R}2$ and $\mathbb{R}3$ on both $G$-MNIST and $G$-ModelNet10 datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (59)
  1. Edward H Adelson and James R Bergen “Spatiotemporal energy models for the perception of motion” In Josa a 2.2 Optical Society of America, 1985, pp. 284–299
  2. Kenneth Atz, Francesca Grisoni and Gisbert Schneider “Geometric deep learning on molecular representations” In Nature Machine Intelligence 3.12 Nature Publishing Group UK London, 2021, pp. 1023–1032
  3. Florentin Bieder, Robin Sandkuhler and Philippe C Cattin “Comparison of methods generalizing max-and average-pooling” In arXiv preprint arXiv:2103.01746, 2021
  4. “Approximating cnns with bag-of-local-features models works surprisingly well on imagenet” In arXiv preprint arXiv:1904.00760, 2019
  5. D Brillinger “Some history of higher-order statistics and spectra” In Stat. Sin. 1, 1991, pp. 465–476
  6. “Geometric Deep Learning: Grids, Groups, Graphs, Geodesics, and Gauges” In arXiv preprint arXiv:2104.13478, 2021
  7. Gabriele Cesa, Leon Lang and Maurice Weiler “A Program to Build E(N)-Equivariant Steerable CNNs” In International Conference on Learning Representations, 2022 URL: https://openreview.net/forum?id=WE4qe9xlnQw
  8. “Group equivariant convolutional networks” In International conference on machine learning, 2016, pp. 2990–2999 PMLR
  9. Taco S Cohen and Max Welling “Steerable cnns” In arXiv preprint arXiv:1612.08498, 2016
  10. Taco S. Cohen, Mario Geiger and Maurice Weiler “A general theory of equivariant CNNs on homogeneous spaces” In Advances in Neural Information Processing Systems 32.NeurIPS, 2019
  11. “Maximal function pooling with applications” In Excursions in Harmonic Analysis, Volume 6: In Honor of John Benedetto’s 80th Birthday Springer, 2021, pp. 413–429
  12. Anastasios Delopoulos, Andreas Tirakis and Stefanos Kollias “Invariant image classification using triple-correlation-based neural networks” In IEEE Transactions on Neural Networks 5.3 IEEE, 1994, pp. 392–408
  13. Sarita S Deshpande, Graham A Smith and Wim Drongelen “Third-order motifs are sufficient to fully and uniquely characterize spatiotemporal neural network activity” In Scientific Reports 13.1 Nature Publishing Group UK London, 2023, pp. 238
  14. “Alpha-Integration Pooling for Convolutional Neural Networks”, 2020 arXiv:1811.03436 [cs.LG]
  15. Joan Bruna Estrach, Arthur Szlam and Yann LeCun “Signal recovery from pooling representations” In International conference on machine learning, 2014, pp. 307–315 PMLR
  16. “Geometric Lp-norm feature pooling for image classification” In CVPR 2011, 2011, pp. 2609–2704 IEEE
  17. K. Fukushima “Neocognitron: A Self-organizing Neural Network Model for a Mechanism of Pattern Recognition Unaffected by Shift in Position” In Biological Cybernetics 36, 1980, pp. 193–202
  18. “Deciphering interaction fingerprints from protein molecular surfaces using geometric deep learning” In Nature Methods 17.2 Nature Publishing Group US New York, 2020, pp. 184–192
  19. “Global second-order pooling convolutional networks” In Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition, 2019, pp. 3024–3033
  20. Ian J Goodfellow, Jonathon Shlens and Christian Szegedy “Explaining and harnessing adversarial examples” In arXiv preprint arXiv:1412.6572, 2014
  21. Benjamin Graham “Fractional max-pooling” In arXiv preprint arXiv:1412.6071, 2014
  22. “Learned-norm pooling for deep feedforward and recurrent neural networks” In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2014, Nancy, France, September 15-19, 2014. Proceedings, Part I 14, 2014, pp. 530–546 Springer
  23. Brian C Hall “Lie groups, Lie algebras, and representations” In Quantum Theory for Mathematicians Springer, 2013, pp. 333–366
  24. “Spatial pyramid pooling in deep convolutional networks for visual recognition” In IEEE transactions on pattern analysis and machine intelligence 37.9 IEEE, 2015, pp. 1904–1916
  25. David H Hubel and Torsten N Wiesel “Receptive fields of single neurones in the cat’s striate cortex” In The Journal of physiology 148.3 Wiley Online Library, 1959, pp. 574–591
  26. “Excessive invariance causes adversarial vulnerability” In arXiv preprint arXiv:1811.00401, 2018
  27. “Performance of a geometric deep learning pipeline for HL-LHC particle tracking” In The European Physical Journal C 81 Springer, 2021, pp. 1–14
  28. R. Kakarala “A group theoretic approach to the triple correlation” In IEEE Workshop on higher order statistics, 1993, pp. 28–32
  29. Ramakrishna Kakarala “A group-theoretic approach to the triple correlation” In [1993 Proceedings] IEEE Signal Processing Workshop on Higher-Order Statistics, 1993, pp. 28–32 IEEE
  30. Ramakrishna Kakarala “Completeness of bispectrum on compact groups” In arXiv preprint arXiv:0902.0196 1 Citeseer, 2009
  31. Ramakrishna Kakarala “The bispectrum as a source of phase-sensitive invariants for Fourier descriptors: a group-theoretic approach” In Journal of Mathematical Imaging and Vision 44.3 Springer, 2012, pp. 341–353
  32. Ramakrishna Kakarala “Triple correlation on groups”, 1992
  33. Stefanos D Kollias “A multiresolution neural network approach to invariant image recognition” In Neurocomputing 12.1 Elsevier, 1996, pp. 35–57
  34. R. Kondor “Group theoretical methods in machine learning” Columbia University, PhD Thesis, 2008
  35. Ashwani Kumar “Ordinal pooling networks: for preserving information over shrinking feature maps” In arXiv preprint arXiv:1804.02702, 2018
  36. Y LeCun, C Cortes and C Burges “The MNIST Dataset of Handwritten Digits (Images)” In NYU: New York, NY, USA, 1999
  37. Yann LeCun “The MNIST database of handwritten digits” In http://yann. lecun. com/exdb/mnist/, 1998
  38. Tsung-Yu Lin, Aruni RoyChowdhury and Subhransu Maji “Bilinear CNN models for fine-grained visual recognition” In Proceedings of the IEEE international conference on computer vision, 2015, pp. 1449–1457
  39. Chrysostomos L Nikias and Jerry M Mendel “Signal processing with higher-order spectra” In IEEE Signal processing magazine 10.3 IEEE, 1993, pp. 10–37
  40. “Pooling in convolutional neural networks for medical image analysis: a survey and an empirical study” In Neural Computing and Applications 34.7 Springer ScienceBusiness Media LLC, 2022, pp. 5321–5347 DOI: 10.1007/s00521-022-06953-8
  41. “Concentric circle pooling in deep convolutional networks for remote sensing scene classification” In Remote Sensing 10.6 MDPI, 2018, pp. 934
  42. “Bispectral Neural Networks” In International Conference on Learning Representations, 2023
  43. “Multipartite pooling for deep convolutional neural networks” In arXiv preprint arXiv:1710.07435, 2017
  44. Zenglin Shi, Yangdong Ye and Yunpeng Wu “Rank-based pooling for deep convolutional neural networks” In Neural Networks 83 Elsevier, 2016, pp. 21–31
  45. Alexandros Stergiou, Ronald Poppe and Grigorios Kalliatakis “Refining activation downsampling with SoftPool” In Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 10357–10366
  46. Bernd Sturmfels “Algorithms in invariant theory” Springer Science & Business Media, 2008
  47. Zhiqiang Tong, Kazuyuki Aihara and Gouhei Tanaka “A hybrid pooling method for convolutional neural networks” In Neural Information Processing: 23rd International Conference, ICONIP 2016, Kyoto, Japan, October 16–21, 2016, Proceedings, Part II 23, 2016, pp. 454–461 Springer
  48. J. Tukey “The spectral representation and transformation properties of the higher moments of stationary time series” In Reprinted in The Collected Works of John W. Tukey 1, 1953, pp. 165–184
  49. “Building detail-sensitive semantic segmentation networks with polynomial pooling” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 7115–7123
  50. “General E(2)-Equivariant Steerable CNNs” In Conference on Neural Information Processing Systems (NeurIPS), 2019
  51. “Max-pooling dropout for regularization of convolutional neural networks” In Neural Information Processing: 22nd International Conference, ICONIP 2015, Istanbul, Turkey, November 9-12, 2015, Proceedings, Part I 22, 2015, pp. 46–54 Springer
  52. “3d shapenets: A deep representation for volumetric shapes” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 1912–1920
  53. John I Yellott “Implications of triple correlation uniqueness for texture statistics and the Julesz conjecture” In JOSA A 10.5 Optica Publishing Group, 1993, pp. 777–793
  54. JI Yellott Jr and GJ Iverson “Uniqueness theorems for generalized autocorrelations” In Journal of the Optical Society of America 9, 1992, pp. 388–404
  55. “Mixed pooling for convolutional neural networks” In Rough Sets and Knowledge Technology: 9th International Conference, RSKT 2014, Shanghai, China, October 24-26, 2014, Proceedings 9, 2014, pp. 364–375 Springer
  56. “A Comparison of Pooling Methods for Convolutional Neural Networks” In Applied Sciences 12.17 MDPI AG, 2022, pp. 8643 DOI: 10.3390/app12178643
  57. Matthew D Zeiler and Rob Fergus “Stochastic pooling for regularization of deep convolutional neural networks” In arXiv preprint arXiv:1301.3557, 2013
  58. “S3pool: Pooling with stochastic spatial sampling” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 4970–4978
  59. Ning Zhang, Ryan Farrell and Trever Darrell “Pose pooling kernels for sub-category recognition” In 2012 IEEE Conference on Computer Vision and Pattern Recognition, 2012, pp. 3665–3672 IEEE
Citations (2)

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube