Directional diffusion models for graph representation learning (2306.13210v1)
Abstract: In recent years, diffusion models have achieved remarkable success in various domains of artificial intelligence, such as image synthesis, super-resolution, and 3D molecule generation. However, the application of diffusion models in graph learning has received relatively little attention. In this paper, we address this gap by investigating the use of diffusion models for unsupervised graph representation learning. We begin by identifying the anisotropic structures of graphs and a crucial limitation of the vanilla forward diffusion process in learning anisotropic structures. This process relies on continuously adding an isotropic Gaussian noise to the data, which may convert the anisotropic signals to noise too quickly. This rapid conversion hampers the training of denoising neural networks and impedes the acquisition of semantically meaningful representations in the reverse process. To address this challenge, we propose a new class of models called {\it directional diffusion models}. These models incorporate data-dependent, anisotropic, and directional noises in the forward diffusion process. To assess the efficacy of our proposed models, we conduct extensive experiments on 12 publicly available datasets, focusing on two distinct graph representation learning tasks. The experimental results demonstrate the superiority of our models over state-of-the-art baselines, indicating their effectiveness in capturing meaningful graph representations. Our studies not only provide valuable insights into the forward process of diffusion models but also highlight the wide-ranging potential of these models for various graph-related tasks.
- Diffusion-based representation learning. arXiv preprint arXiv:2105.14257.
- Cold diffusion: Inverting arbitrary image transforms without noise. arXiv preprint arXiv:2208.09392.
- Label-efficient semantic segmentation with diffusion models. arXiv preprint arXiv:2112.03126.
- Libsvm: a library for support vector machines. ACM transactions on intelligent systems and technology (TIST), 2(3):1–27.
- Perception prioritized training of diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11472–11481.
- Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems, 34:8780–8794.
- Representation degeneration problem in training natural language generation models. arXiv preprint arXiv:1907.12009.
- Diffusion models for graphs benefit from discrete state spaces. arXiv preprint arXiv:2210.01549.
- Contrastive multi-view representation learning on graphs. In International conference on machine learning, pages 4116–4126. PMLR.
- Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840–6851.
- Graphmae: Self-supervised masked graph autoencoders. arXiv preprint arXiv:2205.10803.
- Open graph benchmark: Datasets for machine learning on graphs. Advances in neural information processing systems, 33:22118–22133.
- Gpt-gnn: Generative pre-training of graph neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 1857–1867.
- Score-based generative modeling of graphs via the system of stochastic differential equations. arXiv preprint arXiv:2202.02514.
- On the sentence embeddings from pre-trained language models. arXiv preprint arXiv:2011.05864.
- Diffusion-lm improves controllable text generation. arXiv preprint arXiv:2205.14217.
- Diffusion autoencoders: Toward a meaningful and decodable representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10619–10629.
- Gcc: Graph contrastive coding for graph neural network pre-training. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pages 1150–1160.
- High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10684–10695.
- Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502.
- Infograph: Unsupervised and semi-supervised graph-level representation learning via mutual information maximization. arXiv preprint arXiv:1908.01000.
- Large-scale representation learning on graphs via bootstrapping. arXiv preprint arXiv:2102.06514.
- Deep graph infomax. ICLR (Poster), 2(3):4.
- Infogcl: Information-aware graph contrastive learning. Advances in Neural Information Processing Systems, 34:30414–30425.
- How powerful are graph neural networks? arXiv preprint arXiv:1810.00826.
- Deep graph kernels. In Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, pages 1365–1374.
- Revisiting semi-supervised learning with graph embeddings. In International conference on machine learning, pages 40–48. PMLR.
- Hierarchical graph representation learning with differentiable pooling. Advances in neural information processing systems, 31.
- Graph contrastive learning automated. In International Conference on Machine Learning, pages 12121–12132. PMLR.
- Graph contrastive learning with augmentations. Advances in neural information processing systems, 33:5812–5823.
- From canonical correlation analysis to self-supervised graph neural networks. Advances in Neural Information Processing Systems, 34:76–89.
- Unsupervised representation learning from pre-trained diffusion probabilistic models. arXiv preprint arXiv:2212.12990.
- Deep graph contrastive representation learning. arXiv preprint arXiv:2006.04131.