Boosting Multitask Learning on Graphs through Higher-Order Task Affinities (2306.14009v4)
Abstract: Predicting node labels on a given graph is a widely studied problem with many applications, including community detection and molecular graph prediction. This paper considers predicting multiple node labeling functions on graphs simultaneously and revisits this problem from a multitask learning perspective. For a concrete example, consider overlapping community detection: each community membership is a binary node classification task. Due to complex overlapping patterns, we find that negative transfer is prevalent when we apply naive multitask learning to multiple community detection, as task relationships are highly nonlinear across different node labeling. To address the challenge, we develop an algorithm to cluster tasks into groups based on a higher-order task affinity measure. We then fit a multitask model on each task group, resulting in a boosting procedure on top of the baseline model. We estimate the higher-order task affinity measure between two tasks as the prediction loss of one task in the presence of another task and a random subset of other tasks. Then, we use spectral clustering on the affinity score matrix to identify task grouping. We design several speedup techniques to compute the higher-order affinity scores efficiently and show that they can predict negative transfers more accurately than pairwise task affinities. We validate our procedure using various community detection and molecular graph prediction data sets, showing favorable results compared with existing methods. Lastly, we provide a theoretical analysis to show that under a planted block model of tasks on graphs, our affinity scores can provably separate tasks into groups.
- Emmanuel Abbe “Community detection and stochastic block models: recent developments” In Journal of Machine Learning Research, 2017
- Nir Ailon, Moses Charikar and Alantha Newman “Aggregating inconsistent information: ranking and clustering” In JACM, 2008
- Reid Andersen and Kevin J Lang “Communities from seed sets” In World Wide Web, 2006
- “Saliency-Regularized Deep Multi-Task Learning” In KDD, 2022
- “A theory of learning from different domains” In Machine learning, 2010
- Shai Ben-David, Johannes Gehrke and Reba Schuller “A theoretical framework for learning from a pool of disparate data sources” In KDD, 2002
- “Exploiting task relatedness for multiple task learning” In Learning Theory and Kernel Machines, 2003, pp. 567–580 Springer
- Austin R Benson, David F Gleich and Jure Leskovec “Higher-order organization of complex networks” In Science, 2016
- Filippo Maria Bianchi, Daniele Grattarola and Cesare Alippi “Spectral clustering with graph neural networks for graph pooling” In ICML, 2020 PMLR
- “Fast unfolding of communities in large networks” In Journal of statistical mechanics: theory and experiment, 2008
- “Scaling graph neural networks with approximate pagerank” In KDD, 2020
- Francesco Bonchi, David Garcia-Soriano and Francesco Gullo “Correlation clustering” In Synthesis Lectures on Data Mining and Knowledge Discovery, 2022
- Leo Breiman “Random forests” In Machine Learning, 2001
- Rich Caruana “Multitask learning” In Machine learning, 1997
- “Scalable graph neural networks via bidirectional propagation” In NeurIPS, 2020
- Zhengdao Chen, Xiang Li and Joan Bruna “Supervised community detection with line graph neural networks” In ICLR, 2019
- Koby Crammer, Michael Kearns and Jennifer Wortman “Learning from Multiple Sources.” In Journal of Machine Learning Research, 2008
- “Correlation clustering in general weighted graphs” In Theoretical Computer Science, 2006
- “Regularized multi-task learning” In KDD, 2004
- “Efficiently identifying task groupings for multi-task learning” In NeurIPS, 2021
- Santo Fortunato “Community detection in graphs” In Physics Reports, 2010
- “Sign: Scalable inception graph neural networks” In arXiv preprint arXiv:2004.11198, 2020
- “Connectivity in random forests and credit networks” In SODA, 2014
- “Graphonomy: Universal human parsing via graph transfer learning” In CVPR, 2019
- “Graph transfer learning” In Knowledge and Information Systems, 2022
- “node2vec: Scalable feature learning for networks” In KDD, 2016
- “Open graph benchmark: Datasets for machine learning on graphs” In NeurIPS, 2020
- “Strategies for pre-training graph neural networks” In ICLR, 2020
- “Datamodels: Predicting predictions from training data” In ICML, 2022
- “Generalization in Graph Neural Networks: Improved PAC-Bayesian Bounds on Graph Diffusion” In AISTATS, 2023
- Haotian Ju, Dongyue Li and Hongyang R Zhang “Robust fine-tuning of deep neural networks with hessian-based generalization guarantees” In ICML, 2022
- Johannes Klicpera, Aleksandar Bojchevski and Stephan Günnemann “Predict then propagate: Graph neural networks meet personalized pagerank” In ICLR, 2019
- Abhishek Kumar and Hal Daume III “Learning task grouping and overlap in multi-task learning” In ICML, 2012
- “Transfer learning for deep learning on graph-structured data” In AAAI, 2017
- Dongyue Li, Huy Nguyen and Hongyang Ryan Zhang “Task Modeling: Approximating Multitask Predictions for Cross-Task Transfer” In NeurIPS 2022 Workshop on Distribution Shifts: Connecting Methods and Applications, 2022
- Dongyue Li, Huy L Nguyen and Hongyang R Zhang “Identification of Negative Transfers in Multitask Learning Using Surrogate Models” In Trans. Mach. Learn. Res., 2023
- Dongyue Li and Hongyang R. Zhang “Improved regularization and robustness for fine-tuning in neural networks” In NeurIPS, 2021
- “Modeling task relationships in multi-task learning with multi-gate mixture-of-experts” In KDD, 2018
- Domenico Mandaglio, Andrea Tagarelli and Francesco Gullo “Correlation clustering with global weight bounds” In ECML PKDD, 2021
- “Tudataset: A collection of benchmark datasets for learning with graphs” In ICML GRL+ workshop, 2020
- Andrew Ng, Michael Jordan and Yair Weiss “On spectral clustering: Analysis and an algorithm” In Advances in neural information processing systems 14, 2001
- Feiping Nie, Zhanxuan Hu and Xuelong Li “Calibrated multi-task learning” In KDD, 2018
- Sinno Jialin Pan and Qiang Yang “A survey on transfer learning” In TKDE, 2010
- “Gcc: Graph contrastive coding for graph neural network pre-training” In KDD, 2020
- “Normalized cuts and image segmentation” In IEEE Transactions on pattern analysis and machine intelligence Ieee, 2000
- “Which tasks should be learned together in multi-task learning?” In ICML, 2020
- “Verse: Versatile graph embeddings from similarity measures” In WWW, 2018
- “Graph Clustering with Graph Neural Networks” In JMLR, 2023
- “The CLRS algorithmic reasoning benchmark” In ICML, 2022
- “Learning universal graph neural network embeddings with aid of transfer learning” In arXiv preprint arXiv:1909.10086, 2019
- “Exploring and predicting transferability across NLP tasks” In EMNLP, 2020
- “Multi-task feature learning for knowledge graph enhanced recommendation” In World Wide Web, 2019
- “Understanding and improving fairness-accuracy trade-offs in multi-task learning” In KDD, 2021
- Joyce Jiyoung Whang, David F Gleich and Inderjit S Dhillon “Overlapping community detection using seed set expansion” In CIKM, 2013
- “On the generalization effects of linear transformations in data augmentation” In ICML, 2020
- Sen Wu, Hongyang R Zhang and Christopher Ré “Understanding and improving information transfer in multi-task learning” In ICLR, 2020
- “Analysis of information transfer from heterogeneous sources via precise high-dimensional asymptotics” In arXiv preprint arXiv:2010.11750, 2021
- “Defining and evaluating network communities based on ground-truth” In KDD Workshop on Mining Data Semantics, 2012
- “Overlapping community detection at scale: a nonnegative matrix factorization approach” In WSDM, 2013
- “Deep multi-task representation learning: A tensor factorisation approach” In ICLR, 2017
- “Local higher-order graph clustering” In KDD, 2017
- “Gradient surgery for multi-task learning” In NeurIPS, 2020
- Hongyang Zhang, Huacheng Yu and Ashish Goel “Pruning based distance sketches with provable guarantees on random graphs” In The World Wide Web Conference, 2019
- “Graph attention multi-layer perceptron” In KDD, 2022
- “A survey on multi-task learning” In IEEE Transactions on Knowledge and Data Engineering, 2021
- “Transfer learning of graph neural networks with ego-graph information maximization” In NeurIPS, 2021
- Dongyue Li (27 papers)
- Haotian Ju (5 papers)
- Aneesh Sharma (12 papers)
- Hongyang R. Zhang (19 papers)