Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 88 tok/s
Gemini 2.5 Pro 52 tok/s Pro
GPT-5 Medium 12 tok/s Pro
GPT-5 High 19 tok/s Pro
GPT-4o 110 tok/s Pro
GPT OSS 120B 470 tok/s Pro
Kimi K2 197 tok/s Pro
2000 character limit reached

Privately Learning from Graphs with Applications in Fine-tuning Large Language Models (2410.08299v1)

Published 10 Oct 2024 in cs.LG, cs.CL, and cs.CR

Abstract: Graphs offer unique insights into relationships and interactions between entities, complementing data modalities like text, images, and videos. By incorporating relational information from graph data, AI models can extend their capabilities beyond traditional tasks. However, relational data in sensitive domains such as finance and healthcare often contain private information, making privacy preservation crucial. Existing privacy-preserving methods, such as DP-SGD, which rely on gradient decoupling assumptions, are not well-suited for relational learning due to the inherent dependencies between coupled training samples. To address this challenge, we propose a privacy-preserving relational learning pipeline that decouples dependencies in sampled relations during training, ensuring differential privacy through a tailored application of DP-SGD. We apply this method to fine-tune LLMs on sensitive graph data, and tackle the associated computational complexities. Our approach is evaluated on LLMs of varying sizes (e.g., BERT, Llama2) using real-world relational data from four text-attributed graphs. The results demonstrate significant improvements in relational learning tasks, all while maintaining robust privacy guarantees during training. Additionally, we explore the trade-offs between privacy, utility, and computational efficiency, offering insights into the practical deployment of our approach. Code is available at https://github.com/Graph-COM/PvGaLM.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (106)
  1. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, pp.  308–318, 2016.
  2. Large-scale differentially private BERT. In Findings of the Association for Computational Linguistics: EMNLP 2022, pp.  6481–6491, 2022.
  3. Jax-privacy: Algorithms for privacy-preserving machine learning in jax, 2022. URL http://github.com/google-deepmind/jax_privacy.
  4. Improving the gaussian mechanism for differential privacy: Analytical calibration and optimal denoising. In International Conference on Machine Learning, pp. 394–403. PMLR, 2018.
  5. Privacy amplification by subsampling: Tight analyses via couplings and divergences. In Advances in Neural Information Processing Systems, volume 31, 2018.
  6. Dp-mix: mixup-based data augmentation for differentially private learning. In Advances in Neural Information Processing Systems, volume 36, 2024.
  7. Benchmarking differential privacy and federated learning for bert models. arXiv preprint arXiv:2106.13973, 2021.
  8. LLM2Vec: Large language models are secretly powerful text encoders. In First Conference on Language Modeling, 2024. URL https://openreview.net/forum?id=IW1PR7vEBf.
  9. SciBERT: A pretrained language model for scientific text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp.  3615–3620, 2019.
  10. Higher-order organization of complex networks. Science, 353(6295):163–166, 2016.
  11. Claude Berge. Hypergraphs: combinatorics of finite sets, volume 45. Elsevier, 1984.
  12. A collaborative filtering approach to mitigate the new user cold start problem. Knowledge-based systems, 26:225–238, 2012.
  13. Translating embeddings for modeling multi-relational data. In Advances in Neural Information Processing Systems, volume 26, 2013.
  14. Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pp.  1877–1901, 2020.
  15. Automatic clipping: Differentially private deep learning made easier and stronger. In Advances in Neural Information Processing Systems, volume 36, 2024.
  16. Zero-shot relational learning for multimodal knowledge graphs. arXiv preprint arXiv:2404.06220, 2024.
  17. A simple framework for contrastive learning of visual representations. In International Conference on Machine Learning, pp. 1597–1607. PMLR, 2020.
  18. Node feature extraction by self-supervised multi-scale neighborhood prediction. In International Conference on Learning Representations, 2022.
  19. Differentially private decoupled graph convolutions for multigranular topology protection. In Advances in Neural Information Processing Systems, volume 36, 2024.
  20. Node-level differentially private graph neural networks. arXiv preprint arXiv:2111.15521, 2021.
  21. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp.  4171–4186, 2019.
  22. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021.
  23. Simteg: A frustratingly simple approach improves textual graph learning. arXiv preprint arXiv:2308.02565, 2023.
  24. Cynthia Dwork. Differential privacy. In International colloquium on automata, languages, and programming, pp.  1–12. Springer, 2006.
  25. The algorithmic foundations of differential privacy. Foundations and Trends® in Theoretical Computer Science, 9(3–4):211–407, 2014.
  26. Differentially private graph learning via sensitivity-bounded personalized pagerank. In Advances in Neural Information Processing Systems, volume 35, pp.  22617–22627, 2022.
  27. SimCSE: Simple contrastive learning of sentence embeddings. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp.  6894–6910, 2021.
  28. Leveraging a medical knowledge graph into large language models for diagnosis prediction. arXiv preprint arXiv:2308.14321, 2023.
  29. Ian Goodfellow. Efficient per-example gradient computations. arXiv preprint arXiv:1510.01799, 2015.
  30. Numerical composition of differential privacy. In Advances in Neural Information Processing Systems, volume 34, pp.  11631–11642, 2021.
  31. Dimensionality reduction by learning an invariant mapping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, volume 2, pp.  1735–1742, 2006.
  32. Representation learning on graphs: Methods and applications. IEEE Data Eng. Bull., 40(3):52–74, 2017.
  33. Accurate estimation of the degree distribution of private networks. In 2009 Ninth IEEE International Conference on Data Mining, pp.  169–178. IEEE, 2009.
  34. Learning and evaluating a differentially private pre-trained language model. In Findings of the Association for Computational Linguistics: EMNLP 2021, pp.  1178–1189, 2021.
  35. Lora: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022.
  36. Open graph benchmark: Datasets for machine learning on graphs. In Advances in Neural Information Processing Systems, volume 33, pp.  22118–22133, 2020.
  37. Large language models on graphs: A comprehensive survey. arXiv preprint arXiv:2312.02783, 2023a.
  38. Patton: Language model pretraining on text-rich networks. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp.  7005–7020, 2023b.
  39. Differentially private language models benefit from public pre-training. In Proceedings of the Second Workshop on Privacy in NLP, pp. 39–45, 2020.
  40. Adam: A method for stochastic optimization. In International Conference on Learning Representations, 2015.
  41. Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations, 2017.
  42. Dp-sgd for non-decomposable objective functions. arXiv preprint arXiv:2310.03104, 2023.
  43. Graph computing for financial crime and fraud detection: Trends, challenges and outlook. International Journal of Semantic Computing, 14(04):565–589, 2020.
  44. What is twitter, a social network or a news media? In Proceedings of the 19th International Conference on World Wide Web, pp.  591–600, 2010.
  45. Scaling up differentially private deep learning with fast per-example gradient clipping. Proceedings on Privacy Enhancing Technologies, 2021.
  46. The dynamics of viral marketing. ACM Transactions on the Web (TWEB), 1(1):5–es, 2007.
  47. Dpcl: Contrastive representation learning with differential privacy. International Journal of Intelligent Systems, 37(11):9701–9725, 2022.
  48. BeLLM: Backward dependency enhanced large language model for sentence embeddings. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pp.  792–804, 2024.
  49. Graphadapter: Tuning vision-language models with dual knowledge graph. In Advances in Neural Information Processing Systems, volume 36, 2023a.
  50. Large language models can be strong differentially private learners. In International Conference on Learning Representations, 2021.
  51. Private graph data release: A survey. ACM Computing Surveys, 55(11):1–39, 2023b.
  52. Learning entity and relation embeddings for knowledge graph completion. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 29, 2015.
  53. Beyond one-model-fits-all: A survey of domain specialization for large language models. arXiv preprint arXiv:2305.18703, 2023.
  54. Disease prediction using graph machine learning based on electronic health data: a review of approaches and trends. In Healthcare, volume 11, pp.  1031. MDPI, 2023.
  55. Foundation models for video understanding: A survey. arXiv preprint arXiv:2405.03770, 2024.
  56. A dataset of networks of computing hosts. In Proceedings of the 2022 ACM on International Workshop on Security and Privacy Analytics, pp.  100–104, 2022.
  57. Image-based recommendations on styles and substitutes. In Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp.  43–52, 2015.
  58. A general approach to adding differential privacy to iterative training procedures. arXiv preprint arXiv:1812.06210, 2018.
  59. Network motifs: simple building blocks of complex networks. Science, 298(5594):824–827, 2002.
  60. Differentially private graph neural networks for whole-graph classification. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(6):7308–7318, 2022.
  61. Releasing graph neural networks with differential privacy guarantees. Transactions on Machine Learning Research, 2023.
  62. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
  63. Modal-adaptive knowledge-enhanced graph-based financial prediction from monetary policy conference calls with llm. arXiv preprint arXiv:2403.16055, 2024.
  64. Training language models to follow instructions with human feedback. In Advances in Neural Information Processing Systems, volume 35, pp.  27730–27744, 2022.
  65. Differential privacy in deep learning: A literature survey. Neurocomputing, pp.  127663, 2024.
  66. ecellm: Generalizing large language models for e-commerce from large-scale, high-quality instruction data. In International Conference on Machine Learning. PMLR, 2024.
  67. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp.  701–710, 2014.
  68. How to dp-fy ml: A practical guide to machine learning with differential privacy. Journal of Artificial Intelligence Research, 77:1113–1201, 2023.
  69. Openalex: A fully-open index of scholarly works, authors, venues, institutions, and concepts. arXiv preprint arXiv:2205.01833, 2022.
  70. Subsampling is not magic: Why large batch sizes work for differentially private stochastic optimisation. In International Conference on Machine Learning. PMLR, 2024.
  71. Lipschitz extensions for node-private graph statistics and the generalized exponential mechanism. In 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS), pp.  495–504. IEEE, 2016.
  72. Locally private graph neural networks. In Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security, pp.  2130–2145, 2021.
  73. Progap: Progressive graph neural networks with differential privacy guarantees. In Proceedings of the 17th ACM International Conference on Web Search and Data Mining, pp.  596–605, 2024.
  74. {{\{{GAP}}\}}: Differentially private graph neural networks with aggregation perturbation. In 32nd USENIX Security Symposium (USENIX Security 23), pp. 3223–3240, 2023.
  75. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.  815–823, 2015.
  76. One size does not fit all: Investigating strategies for differentially-private learning across NLP tasks. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp.  7340–7353, 2022.
  77. Chartalist: Labeled graph datasets for utxo and account-based blockchains. In Advances in Neural Information Processing Systems, volume 35, pp.  34926–34939, 2022.
  78. An overview of microsoft academic service (mas) and applications. In Proceedings of the 24th International Conference on World Wide Web, pp.  243–246, 2015.
  79. Kihyuk Sohn. Improved deep metric learning with multi-class n-pair loss objective. In Advances in Neural Information Processing Systems, volume 29, 2016.
  80. Deep metric learning via lifted structured feature embedding. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016.
  81. Stochastic gradient descent with differentially private updates. In 2013 IEEE Global Conference on Signal and Information Processing, pp.  245–248. IEEE, 2013.
  82. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
  83. Differentially private learning needs better features (or much more data). In International Conference on Learning Representations, 2021.
  84. Attention is all you need. Advances in Neural Information Processing Systems, 30, 2017.
  85. Subsampled rényi differential privacy and analytical moments accountant. In The 22nd International Conference on Artificial Intelligence and Statistics, pp.  1226–1235. PMLR, 2019.
  86. Knowledge graph embedding by translating on hyperplanes. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 28, 2014.
  87. Differentially private graph diffusion with applications in personalized pageranks. In Advances in Neural Information Processing Systems, 2024.
  88. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:2303.17564, 2023.
  89. Leveraging graph-based hierarchical medical entity embedding for healthcare applications. Scientific reports, 11(1):5858, 2021.
  90. Graph-aware language model pre-training on a large graph corpus can help multiple graph applications. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp.  5270–5281, 2023.
  91. Dpne: Differentially private network embedding. In Advances in Knowledge Discovery and Data Mining: 22nd Pacific-Asia Conference, pp.  235–246. Springer, 2018.
  92. Privacy-preserving machine learning: Methods, challenges and directions. arXiv preprint arXiv:2108.04417, 2021.
  93. Embedding entities and relations for learning and inference in knowledge bases. In International Conference on Learning Representations, 2015.
  94. Does negative sampling matter? a review with insights into its theory and applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024.
  95. A survey on large language model (llm) security and privacy: The good, the bad, and the ugly. High-Confidence Computing, pp.  100211, 2024.
  96. Deep bidirectional language-knowledge graph pretraining. In Advances in Neural Information Processing Systems, volume 35, pp.  37309–37323, 2022a.
  97. LinkBERT: Pretraining language models with document links. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp.  8003–8016, 2022b.
  98. Graph convolutional neural networks for web-scale recommender systems. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp.  974–983, 2018.
  99. Graph contrastive learning with augmentations. In Advances in Neural Information Processing Systems, volume 33, pp.  5812–5823, 2020.
  100. Opacus: User-friendly differential privacy library in pytorch. arXiv preprint arXiv:2109.12298, 2021.
  101. Differentially private fine-tuning of language models. In International Conference on Learning Representations, 2022.
  102. Vision-language models for vision tasks: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024.
  103. Graph embedding matrix sharing with differential privacy. IEEE Access, 7:89390–89399, 2019.
  104. Greaselm: Graph reasoning enhanced language models. In International Conference on Learning Representations, 2022.
  105. What you like, what i am: online dating recommendation via matching individual preferences with features. IEEE Transactions on Knowledge and Data Engineering, 35(5):5400–5412, 2022.
  106. Touchup-g: Improving feature representation through graph-centric finetuning. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp.  2662–2666, 2024.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.