Hyperbolic Convolution via Kernel Point Aggregation (2306.08862v1)
Abstract: Learning representations according to the underlying geometry is of vital importance for non-Euclidean data. Studies have revealed that the hyperbolic space can effectively embed hierarchical or tree-like data. In particular, the few past years have witnessed a rapid development of hyperbolic neural networks. However, it is challenging to learn good hyperbolic representations since common Euclidean neural operations, such as convolution, do not extend to the hyperbolic space. Most hyperbolic neural networks do not embrace the convolution operation and ignore local patterns. Others either only use non-hyperbolic convolution, or miss essential properties such as equivariance to permutation. We propose HKConv, a novel trainable hyperbolic convolution which first correlates trainable local hyperbolic features with fixed kernel points placed in the hyperbolic space, then aggregates the output features within a local neighborhood. HKConv not only expressively learns local features according to the hyperbolic geometry, but also enjoys equivariance to permutation of hyperbolic points and invariance to parallel transport of a local neighborhood. We show that neural networks with HKConv layers advance state-of-the-art in various tasks.
- R. Sonthalia and A. Gilbert, “Tree! i am no tree! i am a low dimensional hyperbolic embedding,” Advances in Neural Information Processing Systems, vol. 33, pp. 845–856, 2020.
- Y. Nagano, S. Yamaguchi, Y. Fujita, and M. Koyama, “A wrapped normal distribution on hyperbolic space for gradient-based learning,” in International Conference on Machine Learning. PMLR, 2019, pp. 4693–4702.
- I. Chami, A. Wolf, D.-C. Juan, F. Sala, S. Ravi, and C. Ré, “Low-dimensional hyperbolic knowledge graph embeddings,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Online: Association for Computational Linguistics, Jul. 2020, pp. 6901–6914. [Online]. Available: https://aclanthology.org/2020.acl-main.617
- M. G. Atigh, J. Schoep, E. Acar, N. van Noord, and P. Mettes, “Hyperbolic image segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2022, pp. 4453–4462.
- O. Ganea, G. Bécigneul, and T. Hofmann, “Hyperbolic neural networks,” Advances in neural information processing systems, vol. 31, pp. 5345–5355, 2018.
- I. Chami, Z. Ying, C. Ré, and J. Leskovec, “Hyperbolic graph convolutional neural networks,” Advances in neural information processing systems, vol. 32, pp. 4868–4879, 2019.
- R. Shimizu, Y. Mukuta, and T. Harada, “Hyperbolic neural networks++,” in International Conference on Learning Representations, 2021.
- W. Chen, X. Han, Y. Lin, H. Zhao, Z. Liu, P. Li, M. Sun, and J. Zhou, “Fully hyperbolic neural networks,” in Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022, pp. 5672–5686.
- Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” nature, vol. 521, no. 7553, pp. 436–444, 2015.
- V. Khrulkov, L. Mirvakhabova, E. Ustinova, I. Oseledets, and V. Lempitsky, “Hyperbolic image embeddings,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 6418–6428.
- O. Ahmad and F. Lecue, “FisheyeHDK: Hyperbolic deformable kernel learning for ultra-wide field-of-view image recognition,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, no. 6, pp. 5968–5975, Jun. 2022. [Online]. Available: https://ojs.aaai.org/index.php/AAAI/article/view/20542
- Q. Liu, M. Nickel, and D. Kiela, “Hyperbolic graph neural networks,” Advances in Neural Information Processing Systems, vol. 32, pp. 8230–8241, 2019.
- G. Bachmann, G. Bécigneul, and O. Ganea, “Constant curvature graph convolutional networks,” in International Conference on Machine Learning. PMLR, 2020, pp. 486–496.
- J. Dai, Y. Wu, Z. Gao, and Y. Jia, “A hyperbolic-to-hyperbolic graph convolutional network,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 154–163.
- Y. Zhang, X. Wang, C. Shi, N. Liu, and G. Song, “Lorentzian graph convolutional networks,” in Proceedings of the Web Conference 2021, 2021, pp. 1249–1261.
- M. Atzmon, H. Maron, and Y. Lipman, “Point convolutional neural networks by extension operators,” ACM Trans. Graph., vol. 37, no. 4, jul 2018. [Online]. Available: https://doi.org/10.1145/3197517.3201301
- H. Thomas, C. R. Qi, J.-E. Deschaud, B. Marcotegui, F. Goulette, and L. J. Guibas, “Kpconv: Flexible and deformable convolution for point clouds,” in Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 6411–6420.
- Z.-H. Lin, S.-Y. Huang, and Y.-C. F. Wang, “Convolution in the cloud: Learning deformable kernels in 3d graph convolution networks for point cloud analysis,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 1800–1809.
- J. W. Cannon, W. J. Floyd, R. Kenyon, W. R. Parry et al., “Hyperbolic geometry,” Flavors of geometry, vol. 31, no. 59-115, p. 2, 1997.
- M. Nickel and D. Kiela, “Learning continuous hierarchies in the lorentz model of hyperbolic geometry,” in International Conference on Machine Learning. PMLR, 2018, pp. 3779–3788.
- M. Law, R. Liao, J. Snell, and R. Zemel, “Lorentzian distance learning for hyperbolic representations,” in International Conference on Machine Learning. PMLR, 2019, pp. 3672–3681.
- G. Becigneul and O.-E. Ganea, “Riemannian adaptive optimization methods,” in International Conference on Learning Representations, 2019. [Online]. Available: https://openreview.net/forum?id=r1eiqi09K7
- C. Gulcehre, M. Denil, M. Malinowski, A. Razavi, R. Pascanu, K. M. Hermann, P. Battaglia, V. Bapst, D. Raposo, A. Santoro, and N. de Freitas, “Hyperbolic attention networks,” in International Conference on Learning Representations, 2019. [Online]. Available: https://openreview.net/forum?id=rJxHsjRqFQ
- G. Nikolentzos, M. Chatzianastasis, and M. Vazirgiannis, “Weisfeiler and leman go hyperbolic: Learning distance preserving node representations,” in International Conference on Artificial Intelligence and Statistics. PMLR, 2023, pp. 1037–1054.
- T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” in International Conference on Learning Representations, 2017.
- K. Xu, W. Hu, J. Leskovec, and S. Jegelka, “How powerful are graph neural networks?” in International Conference on Learning Representations, 2019. [Online]. Available: https://openreview.net/forum?id=ryGs6iA5Km
- J. Baek, M. Kang, and S. J. Hwang, “Accurate learning of graph representations with graph multiset pooling,” in International Conference on Learning Representations, 2021. [Online]. Available: https://openreview.net/forum?id=JHcqXGaqiGn
- W. Hu, M. Fey, M. Zitnik, Y. Dong, H. Ren, B. Liu, M. Catasta, and J. Leskovec, “Open graph benchmark: Datasets for machine learning on graphs,” arXiv preprint arXiv:2005.00687, 2020.
- C. Helma, R. D. King, S. Kramer, and A. Srinivasan, “The predictive toxicology challenge 2000–2001,” Bioinformatics, vol. 17, no. 1, pp. 107–108, 2001.
- I. Schomburg, A. Chang, C. Ebeling, M. Gremse, C. Heldt, G. Huhn, and D. Schomburg, “Brenda, the enzyme database: updates and major new developments,” Nucleic acids research, vol. 32, no. suppl_1, pp. D431–D433, 2004.
- K. M. Borgwardt, C. S. Ong, S. Schönauer, S. Vishwanathan, A. J. Smola, and H.-P. Kriegel, “Protein function prediction via graph kernels,” Bioinformatics, vol. 21, no. suppl_1, pp. i47–i56, 2005.
- P. Yanardag and S. Vishwanathan, “Deep graph kernels,” in Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, 2015, pp. 1365–1374.
- P. Veličković, G. Cucurull, A. Casanova, A. Romero, P. Liò, and Y. Bengio, “Graph attention networks,” in International Conference on Learning Representations, 2018. [Online]. Available: https://openreview.net/forum?id=rJXMpikCZ
- W. Hamilton, Z. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” Advances in neural information processing systems, vol. 30, 2017.
- M. Chen, Z. Wei, Z. Huang, B. Ding, and Y. Li, “Simple and deep graph convolutional networks,” in International Conference on Machine Learning. PMLR, 2020, pp. 1725–1735.
- H. Pei, B. Wei, K. C.-C. Chang, Y. Lei, and B. Yang, “Geom-gcn: Geometric graph convolutional networks,” in International Conference on Learning Representations, 2019.
- S. Suresh, V. Budde, J. Neville, P. Li, and J. Ma, “Breaking the limit of graph neural networks by improving the assortativity of graphs with local mixing patterns,” in KDD, 2021.
- D. Lim, F. Hohne, X. Li, S. L. Huang, V. Gupta, O. Bhalerao, and S. N. Lim, “Large scale learning on non-homophilous graphs: New benchmarks and strong simple methods,” Advances in Neural Information Processing Systems, vol. 34, pp. 20 887–20 902, 2021.
- X. Li, R. Zhu, Y. Cheng, C. Shan, S. Luo, D. Li, and W. Qian, “Finding global homophily in graph neural networks when meeting heterophily,” in International Conference on Machine Learning. PMLR, 2022, pp. 13 242–13 256.
- M. Craven, D. DiPasquo, D. Freitag, A. McCallum, T. Mitchell, K. Nigam, and S. Slattery, “Learning to construct knowledge bases from the world wide web,” Artificial intelligence, vol. 118, no. 1-2, pp. 69–113, 2000.
- B. Rozemberczki, C. Allen, and R. Sarkar, “Multi-scale attributed node embedding,” Journal of Complex Networks, vol. 9, no. 2, p. cnab014, 2021.
- J. Tang, J. Sun, C. Wang, and Z. Yang, “Social influence analysis in large-scale networks,” in Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining, 2009, pp. 807–816.
- G. Namata, B. London, L. Getoor, B. Huang, and U. Edu, “Query-driven active surveying for collective classification,” in 10th International Workshop on Mining and Learning with Graphs, vol. 8, 2012, p. 1.
- P. Sen, G. Namata, M. Bilgic, L. Getoor, B. Galligher, and T. Eliassi-Rad, “Collective classification in network data,” AI magazine, vol. 29, no. 3, pp. 93–93, 2008.
- J. Gehring, M. Auli, D. Grangier, D. Yarats, and Y. N. Dauphin, “Convolutional sequence to sequence learning,” in International conference on machine learning. PMLR, 2017, pp. 1243–1252.
- M. Ott, S. Edunov, D. Grangier, and M. Auli, “Scaling neural machine translation,” in Proceedings of the Third Conference on Machine Translation, Volume 1: Research Papers. Belgium, Brussels: Association for Computational Linguistics, October 2018, pp. 1–9. [Online]. Available: http://www.aclweb.org/anthology/W18-6301
- Y. Chen, X. Dai, M. Liu, D. Chen, L. Yuan, and Z. Liu, “Dynamic convolution: Attention over convolution kernels,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 11 030–11 039.
- H. Qin, Y. Ding, M. Zhang, Q. YAN, A. Liu, Q. Dang, Z. Liu, and X. Liu, “BiBERT: Accurate fully binarized BERT,” in International Conference on Learning Representations, 2022. [Online]. Available: https://openreview.net/forum?id=5xEgrl_5FAJ
- X. Fan, C.-H. Yang, and B. C. Vemuri, “Nested hyperbolic spaces for dimensionality reduction and hyperbolic nn design,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 356–365.
- W. Peng, T. Varanka, A. Mostafa, H. Shi, and G. Zhao, “Hyperbolic deep neural networks: A survey,” IEEE Transactions on Pattern Analysis & Machine Intelligence, December 2021.
- T. S. Cohen, M. Geiger, J. Köhler, and M. Welling, “Spherical CNNs,” in International Conference on Learning Representations, 2018. [Online]. Available: https://openreview.net/forum?id=Hkbd5xZRb
- R. Hanocka, A. Hertz, N. Fish, R. Giryes, S. Fleishman, and D. Cohen-Or, “Meshcnn: a network with an edge,” ACM Transactions on Graphics (TOG), vol. 38, no. 4, pp. 1–12, 2019.
- T. Cohen, M. Weiler, B. Kicanaoglu, and M. Welling, “Gauge equivariant convolutional networks and the icosahedral cnn,” in International conference on Machine learning. PMLR, 2019, pp. 1321–1330.
- M. Yang, M. Zhou, Z. Li, J. Liu, L. Pan, H. Xiong, and I. King, “Hyperbolic graph neural networks: A review of methods and applications,” 2022.
- K. Lensink, B. Peters, and E. Haber, “Fully hyperbolic convolutional neural networks,” Research in the Mathematical Sciences, vol. 9, no. 4, p. 60, 2022.