Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

EMP: Effective Multidimensional Persistence for Graph Representation Learning (2401.13713v1)

Published 24 Jan 2024 in cs.LG, cs.AI, and cs.CG

Abstract: Topological data analysis (TDA) is gaining prominence across a wide spectrum of machine learning tasks that spans from manifold learning to graph classification. A pivotal technique within TDA is persistent homology (PH), which furnishes an exclusive topological imprint of data by tracing the evolution of latent structures as a scale parameter changes. Present PH tools are confined to analyzing data through a single filter parameter. However, many scenarios necessitate the consideration of multiple relevant parameters to attain finer insights into the data. We address this issue by introducing the Effective Multidimensional Persistence (EMP) framework. This framework empowers the exploration of data by simultaneously varying multiple scale parameters. The framework integrates descriptor functions into the analysis process, yielding a highly expressive data summary. It seamlessly integrates established single PH summaries into multidimensional counterparts like EMP Landscapes, Silhouettes, Images, and Surfaces. These summaries represent data's multidimensional aspects as matrices and arrays, aligning effectively with diverse ML models. We provide theoretical guarantees and stability proofs for EMP summaries. We demonstrate EMP's utility in graph classification tasks, showing its effectiveness. Results reveal that EMP enhances various single PH descriptors, outperforming cutting-edge methods on multiple benchmark datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (53)
  1. Time-conditioned dances with simplicial complexes: Zigzag filtration curve based supra-hodge convolution networks for time-series forecasting. NeurIPS, 35:8940–8953, 2022.
  2. Blockchain networks: Data structures of Bitcoin, Monero, Zcash, Ethereum, Ripple and Iota. arXiv:2103.08712, 2021.
  3. TAMP-S2GCNets: coupling time-aware multipersistence knowledge representation with spatio-supra graph convolutional networks for time-series forecasting. In ICLR, 2021.
  4. An introduction to topological data analysis: fundamental and practical aspects for data scientists. Frontiers in Artificial Intelligence, 4, 2021.
  5. A survey of topological machine learning methods. Frontiers in Artificial Intelligence, 4:52, 2021.
  6. Barbara Giunti. TDA applications library, 2022. https://www.zotero.org/groups/2425412/tda-applications/library.
  7. An introduction to multiparameter persistence. arXiv preprint arXiv:2203.14289, 2022.
  8. Interactive visualization of 2-D persistence modules. arXiv preprint arXiv:1512.00180, 2015.
  9. Multiparameter persistence image for topological machine learning. NeurIPS, 33, 2020.
  10. Signed barcodes for multi-parameter persistence via rank decompositions. In 38th International Symposium on Computational Geometry (SoCG 2022). Schloss Dagstuhl-Leibniz-Zentrum für Informatik, 2022.
  11. Oliver Vipond. Multiparameter persistence landscapes. J. Mach. Learn. Res., 21:61–1, 2020.
  12. A framework for fast and stable representations of multiparameter persistent homology decompositions. NeurIPS, 2023a.
  13. Stable vectorization of multiparameter persistent homology using signed barcodes as measures. NeurIPS, 2023b.
  14. Spectral networks and locally connected networks on graphs. In ICLR, 2014.
  15. Convolutional neural networks on graphs with fast localized spectral filtering. In NeurIPS, pages 3844–3852, 2016.
  16. Semi-supervised classification with graph convolutional networks. ICLR, 2017.
  17. Predict then propagate: Graph neural networks meet personalized pagerank. In ICLR, 2019.
  18. Mixhop: Higher-order graph convolutional architectures via sparsified neighborhood mixing. In ICML, 2019.
  19. N-GCN: Multi-scale graph convolution for semi-supervised node classification. In Uncertainty in Artificial Intelligence, pages 841–851, 2020.
  20. LFGCN: Levitating over graphs with levy flights. In IEEE ICDM, 2020.
  21. MotifNet: a motif-based graph convolutional network for directed graphs. In IEEE DSW, pages 225–228, 2018.
  22. Graph attention networks. ICLR, 2018.
  23. Spagan: shortest path graph attention network. In International Joint Conference on Artificial Intelligence, pages 4099–4105, 2019.
  24. Gmnn: Graph markov neural networks. In ICML, pages 5241–5250. PMLR, 2019.
  25. Computational Topology for Data Analysis. Cambridge University Press, 2022.
  26. A survey on graph kernels. Applied Network Science, 5(1):1–42, 2020.
  27. A survey of vectorization methods in topological data analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  28. Persistence images: A stable vector representation of persistent homology. Journal of Machine Learning Research, 18, 2017.
  29. P. Bubenik. Statistical topological data analysis using persistence landscapes. JMLR, 16(1):77–102, 2015.
  30. Stochastic convergence of persistence landscapes and silhouettes. In Proceedings of the thirtieth annual symposium on Computational geometry, pages 474–483, 2014.
  31. Persistence curves: A canonical framework for summarizing persistence diagrams. Advances in Computational Mathematics, 48(1):6, 2022.
  32. ToDD: Topological compound fingerprinting in computer-aided drug discovery. NeurIPS, 35:27978–27993, 2022.
  33. Computing minimal presentations and bigraded Betti numbers of 2-parameter persistent homology. SIAM Journal on Applied Algebra and Geometry, 6(2):267–298, 2022.
  34. Spline-fitting with a genetic algorithm: A method for developing classification structure-activity relationships. Journal of chemical information and computer sciences, 43(6):1906–1915, 2003.
  35. Subgraph matching kernels for attributed graphs. In ICML, pages 291–298, 2012.
  36. Protein function prediction via graph kernels. Bioinformatics, 21(suppl_1):i47–i56, 2005.
  37. Faster kernels for graphs with continuous attributes via hashing. In 2016 IEEE 16th International Conference on Data Mining (ICDM), pages 1095–1100. IEEE, 2016.
  38. The multiscale laplacian graph kernel. NeurIPS, 29, 2016.
  39. Weisfeiler-lehman graph kernels. Journal of Machine Learning Research, 12(9), 2011.
  40. On valid optimal assignment kernels and applications to graph classification. NeurIPS, 29, 2016.
  41. Filtration curves for graph representation. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pages 1267–1275, 2021.
  42. A fair comparison of graph neural networks for graph classification. In ICLR, 2020.
  43. A kernel for multi-parameter persistent homology. Computers & graphics: X, 2:100005, 2019.
  44. Gril: A 2222-parameter persistence based vectorization for machine learning. arXiv preprint arXiv:2304.04970, 2023.
  45. On the stability of multigraded Betti numbers and hilbert functions. arXiv preprint arXiv:2112.11901, 2021.
  46. A roadmap for the computation of persistent homology. EPJ Data Science, 6:1–38, 2017.
  47. On the computational complexity of Betti numbers: reductions from matrix rank. In Proceedings of the twenty-fifth annual ACM-SIAM symposium on discrete algorithms, pages 152–160. SIAM, 2014.
  48. Persistence homology of networks: methods and applications. Applied Network Science, 4(1):1–28, 2019.
  49. Computational topology: an introduction. American Mathematical Soc., 2010.
  50. David Eisenbud. Commutative algebra: with a view toward algebraic geometry, volume 150. Springer Science & Business Media, 2013.
  51. Michael Lesnick. The theory of the interleaving distance on multidimensional persistence modules. Foundations of Computational Mathematics, 15(3):613–650, 2015.
  52. On the stability of persistent entropy and new summary functions for topological data analysis. Pattern Recognition, 107:107509, 2020.
  53. Instability of the Betti sequence for persistent homology and a stabilized version of the betti sequence. Journal of the Korean Society for Industrial and Applied Mathematics, 25(4):296–311, 2021.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets