Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LightHGNN: Distilling Hypergraph Neural Networks into MLPs for $100\times$ Faster Inference (2402.04296v2)

Published 6 Feb 2024 in cs.LG

Abstract: Hypergraph Neural Networks (HGNNs) have recently attracted much attention and exhibited satisfactory performance due to their superiority in high-order correlation modeling. However, it is noticed that the high-order modeling capability of hypergraph also brings increased computation complexity, which hinders its practical industrial deployment. In practice, we find that one key barrier to the efficient deployment of HGNNs is the high-order structural dependencies during inference. In this paper, we propose to bridge the gap between the HGNNs and inference-efficient Multi-Layer Perceptron (MLPs) to eliminate the hypergraph dependency of HGNNs and thus reduce computational complexity as well as improve inference speed. Specifically, we introduce LightHGNN and LightHGNN$+$ for fast inference with low complexity. LightHGNN directly distills the knowledge from teacher HGNNs to student MLPs via soft labels, and LightHGNN$+$ further explicitly injects reliable high-order correlations into the student MLPs to achieve topology-aware distillation and resistance to over-smoothing. Experiments on eight hypergraph datasets demonstrate that even without hypergraph dependency, the proposed LightHGNNs can still achieve competitive or even better performance than HGNNs and outperform vanilla MLPs by $16.3$ on average. Extensive experiments on three graph datasets further show the average best performance of our LightHGNNs compared with all other methods. Experiments on synthetic hypergraphs with 5.5w vertices indicate LightHGNNs can run $100\times$ faster than HGNNs, showcasing their ability for latency-sensitive deployments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (34)
  1. UCI machine learning repository, 2007.
  2. Hypergraph convolution and hypergraph attention. Pattern Recognition, 110:107637, 2021.
  3. A note on over-smoothing for graph neural networks. In 37th International Conference on Machine Learning, 2020.
  4. Preventing over-smoothing for hypergraph neural networks. CoRR, abs/2203.17159, 2022.
  5. You are AllSet: A multiset function framework for hypergraph neural networks. In International Conference on Learning Representations, 2022.
  6. Convolutional neural networks on graphs with fast localized spectral filtering. In Proceedings of the 30th International Conference on Neural Information Processing Systems, pp.  3844–3852, 2016.
  7. HNHN: Hypergraph networks with hyperedge neurons. CoRR, abs/2006.12278, 2020.
  8. Hypergraph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, pp.  3558–3565, 2019.
  9. HpLapGCN: Hypergraph p-laplacian graph convolutional networks. Neurocomputing, 362:166–174, 2019.
  10. MaGNN: Metapath aggregated graph neural network for heterogeneous graph embedding. In Proceedings of The Web Conference 2020, pp.  2331–2341, 2020.
  11. HGNN+: General hypergraph neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(3):3181–3199, 2022.
  12. Citeseer: An automatic citation indexing system. In Proceedings of the third ACM conference on Digital libraries, pp.  89–98, 1998.
  13. Distilling the knowledge in a neural network. stat, 1050:9, 2015.
  14. UniGNN: A unified framework for graph and hypergraph neural networks. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, pp.  2563–2569, 2021.
  15. Dual channel hypergraph collaborative filtering. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pp.  2020–2029, 2020.
  16. Dynamic hypergraph neural networks. In IJCAI, pp.  2635–2641, 2019.
  17. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations, 2017.
  18. Automating the construction of internet portals with machine learning. Information Retrieval, 3:127–163, 2000.
  19. Hypergraph geometry reflects higher-order dynamics in protein interaction networks. Scientific Reports, 12(1):20879, 2022.
  20. Exploring complex and heterogeneous correlations on hypergraph for the prediction of drug-target interactions. Patterns, 2(12), 2021.
  21. HyGNN: Drug-drug interaction prediction via hypergraph neural network. In 2023 IEEE 39th International Conference on Data Engineering (ICDE), pp.  1503–1516, 2023.
  22. Collective classification in network data. AI magazine, 29:93–93, 2008.
  23. Pathsim: Meta path-based top-k similarity search in heterogeneous information networks. Proceedings of the VLDB Endowment, 4(11):992–1003, 2011.
  24. Learning MLPs on graphs: A unified view of effectiveness, robustness, and efficiency. In The Eleventh International Conference on Learning Representations, 2022.
  25. Graph attention networks. stat, 1050(20):10–48550, 2017.
  26. Hypergraph factorization for multi-tissue gene expression imputation. Nature Machine Intelligence, pp.  1–15, 2023.
  27. Equivariant hypergraph diffusion neural operators. In The Eleventh International Conference on Learning Representations, 2023.
  28. Quantifying the knowledge in gnns for reliable distillation into MLPs. In Proceedings of the 40th International Conference on Machine Learning, 2023.
  29. Graph information bottleneck. Advances in Neural Information Processing Systems, 33:20437–20448, 2020.
  30. Hypergraph contrastive collaborative filtering. In Proceedings of the 45th International ACM SIGIR conference on research and development in information retrieval, pp.  70–79, 2022.
  31. HyperGCN: A new method for training graph convolutional networks on hypergraphs. Advances in neural information processing systems, 32, 2019.
  32. Extract the knowledge of graph neural networks and go beyond it: An effective knowledge distillation framework. In Proceedings of the web conference 2021, pp.  1227–1237, 2021.
  33. Graph-less neural networks: Teaching old MLPs new tricks via distillation. In International Conference on Learning Representations, 2021.
  34. Deep hypergraph structure learning. CoRR, abs/2208.12547, 2022.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets