A Directional Diffusion Graph Transformer for Recommendation (2404.03326v1)
Abstract: In real-world recommender systems, implicitly collected user feedback, while abundant, often includes noisy false-positive and false-negative interactions. The possible misinterpretations of the user-item interactions pose a significant challenge for traditional graph neural recommenders. These approaches aggregate the users' or items' neighbours based on implicit user-item interactions in order to accurately capture the users' profiles. To account for and model possible noise in the users' interactions in graph neural recommenders, we propose a novel Diffusion Graph Transformer (DiffGT) model for top-k recommendation. Our DiffGT model employs a diffusion process, which includes a forward phase for gradually introducing noise to implicit interactions, followed by a reverse process to iteratively refine the representations of the users' hidden preferences (i.e., a denoising process). In our proposed approach, given the inherent anisotropic structure observed in the user-item interaction graph, we specifically use anisotropic and directional Gaussian noises in the forward diffusion process. Our approach differs from the sole use of isotropic Gaussian noises in existing diffusion models. In the reverse diffusion process, to reverse the effect of noise added earlier and recover the true users' preferences, we integrate a graph transformer architecture with a linear attention module to denoise the noisy user/item embeddings in an effective and efficient manner. In addition, such a reverse diffusion process is further guided by personalised information (e.g., interacted items) to enable the accurate estimation of the users' preferences on items. Our extensive experiments conclusively demonstrate the superiority of our proposed graph diffusion model over ten existing state-of-the-art approaches across three benchmark datasets.
- How expressive are graph neural networks in recommendation. In Proc. of CIKM.
- D4Explainer: In-distribution GNN explanations via discrete denoising diffusion. arXiv preprint arXiv:2310.19321 (2023).
- Yifan Chen and Maarten de Rijke. 2018. A collective variational autoencoder for top-n recommendation with side information. In Proc. of DLRS.
- Prafulla Dhariwal and Alexander Nichol. 2021. Diffusion models beat gans on image synthesis. In Proc. of NeurIPS.
- Diffuser: efficient transformers with multi-hop attention diffusion for long sequences. In Proc. of AAAI.
- Graph neural networks for recommender system. In Proc. of WSDM.
- Diffusion improves graph learning. In Proc. of NeurIPS.
- Diffusion models for graphs benefit from discrete state spaces. In Proc. of LoG.
- LightGCN: Simplifying and powering graph convolution network for recommendation. In Proc. of SIGIR.
- Session-based recommendations with recurrent neural networks. In Proc. of ICLR.
- Denoising diffusion probabilistic models. In Proc. of NeurIPS.
- MUDiff: Unified diffusion for complete molecule generation. arXiv preprint arXiv:2304.14621 (2023).
- Modeling user preferences in recommender systems: A classification framework for explicit and implicit user feedback. Transactions on Interactive Intelligent Systems (2014).
- Adaptive graph contrastive learning for recommendation. In Proc. of SIGKDD.
- Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In Proc. of ICDM.
- Autoregressive diffusion model for graph generation. In Proc. of ICML.
- Graph transformer for recommendation. Proc. of SIGIR (2023).
- DiffuRec: A Diffusion Model for Sequential Recommendation. arXiv preprint arXiv:2304.00686 (2023).
- Variational autoencoders for collaborative filtering. In Proc. of WebConf.
- Diffusion augmentation for sequential recommendation. In Proc. of CIKM.
- Recommender systems with heterogeneous side information. In Proc. of WebConf.
- Image-based recommendations on styles and substitutes. In Proc. of SIGIR.
- Graph neural pre-training for enhancing recommendations using side information. Transitions on Information Systems (2021).
- Neural anisotropy directions. In Proc. of NeurIPS.
- Are my deep learning systems fair? An empirical study of fixed-seed training. In Proc. of NeuIPS.
- High-resolution image synthesis with latent diffusion models. In Proc. of the CVPR.
- Donald J Schuirmann. 1987. A comparison of the two one-sided tests procedure and the power approach for assessing the equivalence of average bioavailability. Journal of pharmacokinetics and biopharmaceutics 15 (1987).
- Recvae: A new variational autoencoder for top-n recommendations with implicit feedback. In Proc. of WSDM.
- Deep unsupervised learning using nonequilibrium thermodynamics. In Proc. of ICML.
- BERT4Rec: Sequential recommendation with bidirectional encoder representations from transformer. In Proc. of CIKM.
- Cross-domain action recognition via collective matrix factorization with graph Laplacian regularization. Image and Vision Computing 55 (2016), 119–126.
- Attention is all you need. Proc. of NeurIPS (2017).
- DiGress: Discrete denoising diffusion for graph generation. In Proc. of ICLR.
- Recommendation via collaborative diffusion generative model. In Proc. of KSEM.
- Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768 (2020).
- Denoising implicit feedback for recommendation. In Proc. of WSDM. 373–381.
- Diffusion Recommender Model. Proc. of SIGIR.
- KGAT: Knowledge graph attention network for recommendation. In Proc. of SIGKDD.
- Neural graph collaborative filtering. In Proc. of SIGIR.
- Learning intents behind interactions with knowledge graph for recommendation. In Proc. of WWW.
- Max Welling and Thomas N Kipf. 2016. Semi-supervised classification with graph convolutional networks. In Proc. of ICLR.
- Mind: A large-scale dataset for news recommendation. In Proc. of ACL.
- Self-supervised graph learning for recommendation. In Proc. of SIGIR.
- Difformer: Scalable (graph) transformers induced by energy constrained diffusion. arXiv preprint arXiv:2301.09474 (2023).
- Collaborative denoising auto-encoders for top-n recommender systems. In Proc. of WSDM.
- Linear discriminant analysis. Robust data mining (2013), 27–33.
- Directional diffusion models for graph representation learning. Proc. of NeurIPS (2023).
- Knowledge graph contrastive learning for recommendation. In Proc. of SIGIR.
- Towards robust neural graph collaborative filtering via structure denoising and embedding perturbation. Transactions on Information Systems (2023).
- Large multi-modal encoders for recommendation. arXiv preprint arXiv:2310.20343 (2023).
- Contrastive Graph Learning with Positional Representation for Recommendation. (2023).
- Contrastive graph prompt-tuning for cross-domain recommendation. ACM Transactions on Information Systems 42, 2 (2023), 1–28.
- Graph contrastive learning with positional representation for recommendation. In European Conference on Information Retrieval. Springer, 288–303.
- Multi-modal graph contrastive learning for micro-video recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 1807–1811.
- Are graph augmentations necessary? Simple graph contrastive learning for recommendation. In Proc. of SIGIR.
- Self-supervised learning for recommender systems: a survey. In Proc. of SIGIR.
- Collaborative knowledge base embedding for recommender systems. In Proc. of SIGKDD.
- SLED: Structure learning based denoising for recommendation. Transactions on Information Systems (2023).
- S3-rec: Self-supervised learning for sequential recommendation with mutual information maximization. In Proc. of CIKM.
- Multi-level cross-view contrastive learning for knowledge-aware recommender system. In Proc. of SIGIR.
- Diffusion models in nlp: A survey. arXiv preprint arXiv:2305.14671 (2023).