Multi-task Item-attribute Graph Pre-training for Strict Cold-start Item Recommendation (2306.14462v1)
Abstract: Recommendation systems suffer in the strict cold-start (SCS) scenario, where the user-item interactions are entirely unavailable. The ID-based approaches completely fail to work. Cold-start recommenders, on the other hand, leverage item contents to map the new items to the existing ones. However, the existing SCS recommenders explore item contents in coarse-grained manners that introduce noise or information loss. Moreover, informative data sources other than item contents, such as users' purchase sequences and review texts, are ignored. We explore the role of the fine-grained item attributes in bridging the gaps between the existing and the SCS items and pre-train a knowledgeable item-attribute graph for SCS item recommendation. Our proposed framework, ColdGPT, models item-attribute correlations into an item-attribute graph by extracting fine-grained attributes from item contents. ColdGPT then transfers knowledge into the item-attribute graph from various available data sources, i.e., item contents, historical purchase sequences, and review texts of the existing items, via multi-task learning. To facilitate the positive transfer, ColdGPT designs submodules according to the natural forms of the data sources and coordinates the multiple pre-training tasks via unified alignment-and-uniformity losses. Our pre-trained item-attribute graph acts as an implicit, extendable item embedding matrix, which enables the SCS item embeddings to be easily acquired by inserting these items and propagating their attributes' embeddings. We carefully process three public datasets, i.e., Yelp, Amazon-home, and Amazon-sports, to guarantee the SCS setting for evaluation. Extensive experiments show that ColdGPT consistently outperforms the existing SCS recommenders by large margins and even surpasses models that are pre-trained on 75-224 times more, cross-domain data on two out of four datasets.
- CB2CF: a neural multiview content-to-collaborative filtering model for completely cold item recommendations. In Proc. RecSys.
- Language models are few-shot learners. In Proc. NeurIPS.
- Unifying knowledge graph learning and recommendation: Towards a better understanding of user preferences. In Proc. TheWebConf.
- Rich Caruana. 1997. Multitask learning. Machine learning 28, 1 (1997), 41–75.
- Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks. In Proc. ICML.
- Just pick a sign: Optimizing deep multitask models with gradient sign dropout. Proc. NeurIPS.
- Wide & deep learning for recommender systems. In Proc. DLRS.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
- Mamo: Memory-augmented meta-optimization for cold-start recommendation. In Proc. SIGKDD.
- Metakg: Meta-learning on knowledge graph for cold-start recommendation. IEEE TKDE (2022).
- Graph neural networks for social recommendation. In Proc. TheWebConf.
- Zero-shot Item-based Recommendation via Multi-task Product Knowledge Graph Pre-Training. arXiv preprint arXiv:2305.07633 (2023).
- Recommendation as Language Processing (RLP): A Unified Pretrain, Personalized Prompt & Predict Paradigm (P5). In Proc. RecSys.
- Aditya Grover and Jure Leskovec. 2016. node2vec: Scalable feature learning for networks. In Proc. SIGKDD.
- Michael Gutmann and Aapo Hyvärinen. 2010. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proc. AISTATS.
- Pre-training graph neural networks for cold-start users and items representation. In Proc. WSDM.
- Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proc. SIGIR.
- Towards Universal Sequence Representation Learning for Recommender Systems. In Proc. SIGKDD.
- Strategies for pre-training graph neural networks. Proc. ICLR.
- Gpt-gnn: Generative pre-training of graph neural networks. In Proc. SIGKDD.
- Unsupervised pre-training of graph convolutional networks. In ICLR Workshop: Representation Learning on Graphs and Manifolds.
- Adrián Javaloy and Isabel Valera. 2022. RotoGrad: Gradient Homogenization in Multitask Learning. Proc. ICLR.
- Understanding dimensional collapse in contrastive self-supervised learning. arXiv preprint arXiv:2110.09348 (2021).
- Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In Proc. ICDM.
- Thomas N Kipf and Max Welling. 2017. Semi-supervised classification with graph convolutional networks. In Proc. ICLR.
- In defense of the unitary scalarization for deep multi-task learning. Proc. NeurIPS.
- Melu: Meta-learned user preference estimator for cold-start recommendation. In Proc. SIGKDD.
- From zero-shot learning to cold-start recommendation. In Proc. AAAI.
- Task-adaptive neural process for user cold-start recommendation. In Proc. TheWebConf.
- Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. arXiv preprint arXiv:2107.13586 (2021).
- Graph Neural Pre-Training for Recommendation with Side Information. ACM Trans. Inf. Syst. (dec 2022).
- Leveraging distribution alignment via stein path for cross-domain cold-start recommendation. Proc. NeurIPS.
- Augmenting sequential recommendation with pseudo-prior items via reversely pre-training transformer. In Proc. SIGIR.
- Meta-learning on heterogeneous information networks for cold-start recommendation. In Proc. SIGKDD.
- Learning to pre-train graph neural networks. In Proc. AAAI.
- Distributed representations of words and phrases and their compositionality. Proc. NeurIPS.
- Warm up cold-start advertisements: Improving ctr predictions via learning to learn id embeddings. In Proc. SIGIR.
- Pnmta: A pretrained network modulation and task adaptation approach for user cold-start recommendation. In Proc. TheWebConf.
- Aleksandr Petrov and Craig Macdonald. 2022. A Systematic Review and Replicability Study of BERT4Rec for Sequential Recommendation. In Proc. RecSys.
- Attribute graph neural networks for strict cold start recommendation. IEEE TKDE (2020).
- Gcc: Graph contrastive coding for graph neural network pre-training. In Proc. SIGKDD.
- Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 140 (2020), 1–67.
- Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. In Proc. EMNLP.
- Steffen Rendle. 2010. Factorization machines. In Proc. ICDM. IEEE.
- BPR: Bayesian personalized ranking from implicit feedback. In Proc. UAI.
- Ozan Sener and Vladlen Koltun. 2018. Multi-task learning as multi-objective optimization. Proc. NeurIPS.
- BERT4Rec: Sequential recommendation with bidirectional encoder representations from transformer. In Proc. CIKM.
- Gppt: Graph pre-training and prompt tuning to generalize graph neural networks. In Proc. SIGKDD.
- Line: Large-scale information network embedding. In Proc. TheWebConf.
- Ext5: Towards extreme multi-task scaling for transfer learning. Proc. ICLR.
- A meta-learning perspective on cold-start recommendations for items. In Proc. NeurIPS.
- Attention is all you need. Proc. NeurIPS.
- Graph attention networks. In Proc. ICLR.
- Dropoutnet: Addressing cold start in recommender systems. In Proc. NeurIPS.
- Pre-training graph neural network for cross domain recommendation. In 2021 IEEE Third International Conference on Cognitive Machine Intelligence (CogMI). IEEE, 140–145.
- Towards Representation Alignment and Uniformity in Collaborative Filtering. In Proc. SIGKDD.
- Knowledge graph convolutional networks for recommender systems. In Proc. TheWebConf.
- MetaKRec: Collaborative Meta-Knowledge Enhanced Recommender System. In Proc. BigData. IEEE.
- Tongzhou Wang and Phillip Isola. 2020. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In Proc. ICML.
- Kgat: Knowledge graph attention network for recommendation. In Proc. SIGKDD.
- Neural graph collaborative filtering. In Proc. SIGIR.
- Contrastive learning for cold-start recommendation. In Proc. Multimedia.
- Self-supervised graph learning for recommendation. In Proc. SIGIR.
- Towards open-world recommendation: An inductive model-based collaborative filtering approach. In Proc. ICML.
- Attentional factorization machines: Learning the weight of feature interactions via attention networks. In Proc. IJCAI.
- Do Current Multi-Task Optimization Methods in Deep Learning Even Help? Proc. NeurIPS.
- Consisrec: Enhancing gnn for social recommendation via consistent neighbor aggregation. In Proc. SIGIR.
- Mine your own anatomy: Revisiting medical image segmentation with extremely limited labels. arXiv preprint arXiv:2209.13476 (2022).
- ACTION++: Improving Semi-supervised Medical Image Segmentation with Adaptive Anatomical Contrast. arXiv preprint arXiv:2304.02689 (2023).
- Bootstrapping semi-supervised medical image segmentation with anatomical-aware contrastive distillation. In Proc. IPMI. Springer.
- Rethinking semi-supervised medical image segmentation: A variance-reduction perspective. arXiv preprint arXiv:2302.01735 (2023).
- Gradient surgery for multi-task learning. Proc. NeurIPS.
- Where to Go Next for Recommender Systems? ID- vs. Modality-based recommender models revisied. In Proc. SIGIR.
- Do users rate or review? Boost phrase-level sentiment labeling with review-level sentiment classification. In Proc. SIGIR.
- Improving Item Cold-start Recommendation via Model-agnostic Conditional Variational Autoencoder. In Proc. SIGIR.
- Cold brew: Distilling graph node representations with incomplete or missing neighborhoods. In Proc. ICLR.
- Learning to warm up cold item embeddings for cold-start recommendation with meta scaling and shifting networks. In Proc. SIGIR.
- Recommendation for new users and new items via randomized training and mixture-of-experts transformation. In Proc. SIGIR.
- Yuwei Cao (13 papers)
- Liangwei Yang (46 papers)
- Chen Wang (600 papers)
- Zhiwei Liu (114 papers)
- Hao Peng (291 papers)
- Chenyu You (66 papers)
- Philip S. Yu (592 papers)