Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

LiGNN: Graph Neural Networks at LinkedIn (2402.11139v1)

Published 17 Feb 2024 in cs.LG and cs.AI

Abstract: In this paper, we present LiGNN, a deployed large-scale Graph Neural Networks (GNNs) Framework. We share our insight on developing and deployment of GNNs at large scale at LinkedIn. We present a set of algorithmic improvements to the quality of GNN representation learning including temporal graph architectures with long term losses, effective cold start solutions via graph densification, ID embeddings and multi-hop neighbor sampling. We explain how we built and sped up by 7x our large-scale training on LinkedIn graphs with adaptive sampling of neighbors, grouping and slicing of training data batches, specialized shared-memory queue and local gradient optimization. We summarize our deployment lessons and learnings gathered from A/B test experiments. The techniques presented in this work have contributed to an approximate relative improvements of 1% of Job application hearing back rate, 2% Ads CTR lift, 0.5% of Feed engaged daily active users, 0.2% session lift and 0.1% weekly active user lift from people recommendation. We believe that this work can provide practical solutions and insights for engineers who are interested in applying Graph neural networks at large scale.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (33)
  1. adSformers: Personalization from Short-Term Sequences and Diversity of Representations in Etsy Ads. arXiv preprint arXiv:2302.01255 (2023).
  2. Scaling graph neural networks with approximate pagerank. In KDD.
  3. LANNS: a web-scale approximate nearest neighbor lookup system. arXiv preprint arXiv:2010.09426 (2020).
  4. Mixed precision training. https://doi.org/10.48550/arXiv.1710.03740
  5. Relational Deep Learning: Graph Representation Learning on Relational Databases. arXiv:cs.LG/2312.04615
  6. Grale: Designing Networks for Graph Learning. KDD.
  7. Inductive Representation Learning on Large Graphs. NIPS.
  8. MLPInit: Embarrassingly Simple GNN Training Acceleration with MLP Initialization. arXiv:cs.LG/2210.00102
  9. A Multi-Strategy-Based Pre-Training Method for Cold-Start Recommendation. ACM Trans. Inf. Syst. (2023).
  10. Pre-Training Graph Neural Networks for Cold-Start Users and Items Representation. arXiv:cs.IR/2012.07064
  11. How Powerful are Graph Neural Networks? ICLR.
  12. Alex Krizhevsky. 2014. One weird trick for parallelizing convolutional neural networks. https://doi.org/10.48550/arXiv.1404.5997
  13. Kubernetes. 2014. an open-source system for automating deployment. https://kubernetes.io.
  14. GraphCast: Learning skillful medium-range global weather forecasting.
  15. PyTorch-BigGraph: A Large-scale Graph Embedding System. arXiv:cs.LG/1903.12287
  16. Generating Wikipedia by Summarizing Long Sequences. In International Conference on Learning Representations.
  17. Uncertainty-Aware Consistency Learning for Cold-Start Item Recommendation. In SIGIR.
  18. Que2Search: Fast and Accurate Query and Document Understanding for Search at Facebook. KDD.
  19. Heterogeneous Graph Neural Networks for Large-Scale Bid Keyword Matching. CIKM.
  20. HUGE: Huge Unsupervised Graph Embeddings with TPUs. In KDD.
  21. PinnerFormer: Sequence Modeling for User Representation at Pinterest. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 3702–3712.
  22. Attribute Graph Neural Networks for Strict Cold Start Recommendation. IEEE Transactions on Knowledge and Data Engineering (2022).
  23. Nxtpost: User to post recommendations in facebook groups. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. 3792–3800.
  24. Temporal graph networks for deep learning on dynamic graphs. arXiv preprint arXiv:2006.10637 (2020).
  25. Alex Samylkin. 2022. DeepGNN is a framework for training machine learning models on large scale graph data. https://github.com/microsoft/DeepGNN
  26. Graph Neural Networks for Friend Ranking in Large-Scale Social Platforms. KDD.
  27. Modeling Relational Data with Graph Convolutional Networks. In The Semantic Web.
  28. Foundations and modeling of dynamic networks using dynamic graph neural networks: A survey. IEEE Access 9 (2021), 79143–79168.
  29. Attention is all you need. Advances in neural information processing systems 30 (2017).
  30. FORA: Simple and Effective Approximate Single-Source Personalized PageRank. KDD.
  31. TopPPR: Top-k Personalized PageRank Queries with Precision Guarantees on Large Graphs. SIGMOD.
  32. Graph convolutional neural networks for web-scale recommender systems. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining. 974–983.
  33. Da Zheng and Florian Saupe. 2023. Fast-track graph ML with GraphStorm: A new way to solve problems on enterprise-scale graphs. https://aws.amazon.com/blogs/machine-learning/fast-track-graph-ml-with-graphstorm-a-new-way-to-solve-problems-on-enterprise-scale-graphs/
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (23)
  1. Fedor Borisyuk (13 papers)
  2. Shihai He (3 papers)
  3. Yunbo Ouyang (6 papers)
  4. Morteza Ramezani (5 papers)
  5. Peng Du (28 papers)
  6. Xiaochen Hou (6 papers)
  7. Chengming Jiang (3 papers)
  8. Nitin Pasumarthy (1 paper)
  9. Priya Bannur (2 papers)
  10. Birjodh Tiwana (4 papers)
  11. Ping Liu (93 papers)
  12. Siddharth Dangi (2 papers)
  13. Daqi Sun (3 papers)
  14. Zhoutao Pei (1 paper)
  15. Xiao Shi (9 papers)
  16. Sirou Zhu (6 papers)
  17. Qianqi Shen (2 papers)
  18. Kuang-Hsuan Lee (2 papers)
  19. David Stein (6 papers)
  20. Baolei Li (1 paper)
Citations (7)

Summary

  • The paper introduces LiGNN, a scalable graph neural network framework that cuts training time by 7x through adaptive sampling and Kubernetes integration.
  • It leverages graph densification, multi-hop sampling, and a novel temporal modeling approach to effectively learn from dynamic interactions.
  • The deployment significantly enhances LinkedIn’s recommendation systems, boosting job applications, ad clicks, and overall user engagement.

Deploying Large-Scale Graph Neural Networks at LinkedIn: Insights and Innovations

Introduction

LinkedIn, the world's largest professional network, has successfully integrated Graph Neural Networks (GNNs) at an unprecedented scale with its LiGNN framework. This integration aims to optimize LinkedIn's ecosystem, a complex web of professional connections and interactions among members, companies, and various entities. LiGNN encompasses sophisticated algorithmic advancements and strategic deployment mechanisms that address the unique challenges of applying GNNs in a dynamic, large-scale environment.

Challenges and Technological Advancements

LiGNN's development and deployment at LinkedIn presented numerous challenges, notably GNN training scalability, the integration of diverse entities into a unified graph embedding space, addressing the "cold start" problem, and adapting to the platform's dynamic nature. Through inventive approaches, including adaptive sampling and temporal graph architectures, LiGNN has achieved remarkable efficiency and marked improvements in various recommendation systems on LinkedIn.

GNN Training at Scale

LiGNN's training infrastructure utilizes a Kubernetes-based cluster, integrating Microsoft's DeepGNN graph engine for efficient real-time graph sampling. Implementations such as adaptive neighbor sampling and specialized data processing mechanisms have significantly improved GNN training stability and reduced training times by a factor of 7.

Graph Densification and Multi-Hop Sampling

To mitigate issues of graph sparsity and enhance representation learning, LiGNN employs graph densification techniques and multi-hop sampling strategies. These methods have proven effective in enriching graph representations and accelerating the sampling process, demonstrating notable improvements in various production applications.

Temporal Modeling

The framework introduces a simplified and scalable approach to temporal graph modeling, distinct from conventional temporal GNNs. By modifying neighbor sampling to encapsulate temporal dynamics, and incorporating transformer models for sequence encoding, LiGNN captures time-sensitive interactions, contributing to improved model performance across multiple use cases.

Deployment Lessons and Near-Line Inference Architecture

The deployment of LiGNN has provided critical insights, particularly in addressing feature freshness through a GNN model inference pipeline for near real-time embedding generation. Challenges related to integrating LiGNN with LinkedIn's systems prompted the development of specific Apache Beam components, enhancing the pipeline's efficiency and applicability.

Impact Across LinkedIn

LiGNN's deployment across various domains within LinkedIn, including job and people recommendations, ads, and feed post recommendations, has shown substantial improvements. For instance, LiGNN contributed to increases in job application hearing back rates, ad click-through rates, and overall user engagement metrics. The framework demonstrates the practical value and wide applicability of GNNs in enhancing recommendation systems at scale.

Conclusion

The development and deployment of LiGNN at LinkedIn highlight the potential of graph neural networks in transforming large-scale recommendation systems. By addressing key challenges through innovative approaches to training scalability, cold start solutions, and temporal modeling, LiGNN has achieved significant improvements in various application domains. This work serves as a valuable reference for engineers and researchers interested in applying GNNs at scale, promising further advancements and broader applications in the field of AI.

HackerNews