Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Boosting Multitask Learning on Graphs through Higher-Order Task Affinities (2306.14009v4)

Published 24 Jun 2023 in cs.LG and cs.SI

Abstract: Predicting node labels on a given graph is a widely studied problem with many applications, including community detection and molecular graph prediction. This paper considers predicting multiple node labeling functions on graphs simultaneously and revisits this problem from a multitask learning perspective. For a concrete example, consider overlapping community detection: each community membership is a binary node classification task. Due to complex overlapping patterns, we find that negative transfer is prevalent when we apply naive multitask learning to multiple community detection, as task relationships are highly nonlinear across different node labeling. To address the challenge, we develop an algorithm to cluster tasks into groups based on a higher-order task affinity measure. We then fit a multitask model on each task group, resulting in a boosting procedure on top of the baseline model. We estimate the higher-order task affinity measure between two tasks as the prediction loss of one task in the presence of another task and a random subset of other tasks. Then, we use spectral clustering on the affinity score matrix to identify task grouping. We design several speedup techniques to compute the higher-order affinity scores efficiently and show that they can predict negative transfers more accurately than pairwise task affinities. We validate our procedure using various community detection and molecular graph prediction data sets, showing favorable results compared with existing methods. Lastly, we provide a theoretical analysis to show that under a planted block model of tasks on graphs, our affinity scores can provably separate tasks into groups.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (66)
  1. Emmanuel Abbe “Community detection and stochastic block models: recent developments” In Journal of Machine Learning Research, 2017
  2. Nir Ailon, Moses Charikar and Alantha Newman “Aggregating inconsistent information: ranking and clustering” In JACM, 2008
  3. Reid Andersen and Kevin J Lang “Communities from seed sets” In World Wide Web, 2006
  4. “Saliency-Regularized Deep Multi-Task Learning” In KDD, 2022
  5. “A theory of learning from different domains” In Machine learning, 2010
  6. Shai Ben-David, Johannes Gehrke and Reba Schuller “A theoretical framework for learning from a pool of disparate data sources” In KDD, 2002
  7. “Exploiting task relatedness for multiple task learning” In Learning Theory and Kernel Machines, 2003, pp. 567–580 Springer
  8. Austin R Benson, David F Gleich and Jure Leskovec “Higher-order organization of complex networks” In Science, 2016
  9. Filippo Maria Bianchi, Daniele Grattarola and Cesare Alippi “Spectral clustering with graph neural networks for graph pooling” In ICML, 2020 PMLR
  10. “Fast unfolding of communities in large networks” In Journal of statistical mechanics: theory and experiment, 2008
  11. “Scaling graph neural networks with approximate pagerank” In KDD, 2020
  12. Francesco Bonchi, David Garcia-Soriano and Francesco Gullo “Correlation clustering” In Synthesis Lectures on Data Mining and Knowledge Discovery, 2022
  13. Leo Breiman “Random forests” In Machine Learning, 2001
  14. Rich Caruana “Multitask learning” In Machine learning, 1997
  15. “Scalable graph neural networks via bidirectional propagation” In NeurIPS, 2020
  16. Zhengdao Chen, Xiang Li and Joan Bruna “Supervised community detection with line graph neural networks” In ICLR, 2019
  17. Koby Crammer, Michael Kearns and Jennifer Wortman “Learning from Multiple Sources.” In Journal of Machine Learning Research, 2008
  18. “Correlation clustering in general weighted graphs” In Theoretical Computer Science, 2006
  19. “Regularized multi-task learning” In KDD, 2004
  20. “Efficiently identifying task groupings for multi-task learning” In NeurIPS, 2021
  21. Santo Fortunato “Community detection in graphs” In Physics Reports, 2010
  22. “Sign: Scalable inception graph neural networks” In arXiv preprint arXiv:2004.11198, 2020
  23. “Connectivity in random forests and credit networks” In SODA, 2014
  24. “Graphonomy: Universal human parsing via graph transfer learning” In CVPR, 2019
  25. “Graph transfer learning” In Knowledge and Information Systems, 2022
  26. “node2vec: Scalable feature learning for networks” In KDD, 2016
  27. “Open graph benchmark: Datasets for machine learning on graphs” In NeurIPS, 2020
  28. “Strategies for pre-training graph neural networks” In ICLR, 2020
  29. “Datamodels: Predicting predictions from training data” In ICML, 2022
  30. “Generalization in Graph Neural Networks: Improved PAC-Bayesian Bounds on Graph Diffusion” In AISTATS, 2023
  31. Haotian Ju, Dongyue Li and Hongyang R Zhang “Robust fine-tuning of deep neural networks with hessian-based generalization guarantees” In ICML, 2022
  32. Johannes Klicpera, Aleksandar Bojchevski and Stephan Günnemann “Predict then propagate: Graph neural networks meet personalized pagerank” In ICLR, 2019
  33. Abhishek Kumar and Hal Daume III “Learning task grouping and overlap in multi-task learning” In ICML, 2012
  34. “Transfer learning for deep learning on graph-structured data” In AAAI, 2017
  35. Dongyue Li, Huy Nguyen and Hongyang Ryan Zhang “Task Modeling: Approximating Multitask Predictions for Cross-Task Transfer” In NeurIPS 2022 Workshop on Distribution Shifts: Connecting Methods and Applications, 2022
  36. Dongyue Li, Huy L Nguyen and Hongyang R Zhang “Identification of Negative Transfers in Multitask Learning Using Surrogate Models” In Trans. Mach. Learn. Res., 2023
  37. Dongyue Li and Hongyang R. Zhang “Improved regularization and robustness for fine-tuning in neural networks” In NeurIPS, 2021
  38. “Modeling task relationships in multi-task learning with multi-gate mixture-of-experts” In KDD, 2018
  39. Domenico Mandaglio, Andrea Tagarelli and Francesco Gullo “Correlation clustering with global weight bounds” In ECML PKDD, 2021
  40. “Tudataset: A collection of benchmark datasets for learning with graphs” In ICML GRL+ workshop, 2020
  41. Andrew Ng, Michael Jordan and Yair Weiss “On spectral clustering: Analysis and an algorithm” In Advances in neural information processing systems 14, 2001
  42. Feiping Nie, Zhanxuan Hu and Xuelong Li “Calibrated multi-task learning” In KDD, 2018
  43. Sinno Jialin Pan and Qiang Yang “A survey on transfer learning” In TKDE, 2010
  44. “Gcc: Graph contrastive coding for graph neural network pre-training” In KDD, 2020
  45. “Normalized cuts and image segmentation” In IEEE Transactions on pattern analysis and machine intelligence Ieee, 2000
  46. “Which tasks should be learned together in multi-task learning?” In ICML, 2020
  47. “Verse: Versatile graph embeddings from similarity measures” In WWW, 2018
  48. “Graph Clustering with Graph Neural Networks” In JMLR, 2023
  49. “The CLRS algorithmic reasoning benchmark” In ICML, 2022
  50. “Learning universal graph neural network embeddings with aid of transfer learning” In arXiv preprint arXiv:1909.10086, 2019
  51. “Exploring and predicting transferability across NLP tasks” In EMNLP, 2020
  52. “Multi-task feature learning for knowledge graph enhanced recommendation” In World Wide Web, 2019
  53. “Understanding and improving fairness-accuracy trade-offs in multi-task learning” In KDD, 2021
  54. Joyce Jiyoung Whang, David F Gleich and Inderjit S Dhillon “Overlapping community detection using seed set expansion” In CIKM, 2013
  55. “On the generalization effects of linear transformations in data augmentation” In ICML, 2020
  56. Sen Wu, Hongyang R Zhang and Christopher Ré “Understanding and improving information transfer in multi-task learning” In ICLR, 2020
  57. “Analysis of information transfer from heterogeneous sources via precise high-dimensional asymptotics” In arXiv preprint arXiv:2010.11750, 2021
  58. “Defining and evaluating network communities based on ground-truth” In KDD Workshop on Mining Data Semantics, 2012
  59. “Overlapping community detection at scale: a nonnegative matrix factorization approach” In WSDM, 2013
  60. “Deep multi-task representation learning: A tensor factorisation approach” In ICLR, 2017
  61. “Local higher-order graph clustering” In KDD, 2017
  62. “Gradient surgery for multi-task learning” In NeurIPS, 2020
  63. Hongyang Zhang, Huacheng Yu and Ashish Goel “Pruning based distance sketches with provable guarantees on random graphs” In The World Wide Web Conference, 2019
  64. “Graph attention multi-layer perceptron” In KDD, 2022
  65. “A survey on multi-task learning” In IEEE Transactions on Knowledge and Data Engineering, 2021
  66. “Transfer learning of graph neural networks with ego-graph information maximization” In NeurIPS, 2021
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Dongyue Li (27 papers)
  2. Haotian Ju (5 papers)
  3. Aneesh Sharma (12 papers)
  4. Hongyang R. Zhang (19 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.