Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Large Language Models meet Collaborative Filtering: An Efficient All-round LLM-based Recommender System (2404.11343v2)

Published 17 Apr 2024 in cs.IR and cs.AI
Large Language Models meet Collaborative Filtering: An Efficient All-round LLM-based Recommender System

Abstract: Collaborative filtering recommender systems (CF-RecSys) have shown successive results in enhancing the user experience on social media and e-commerce platforms. However, as CF-RecSys struggles under cold scenarios with sparse user-item interactions, recent strategies have focused on leveraging modality information of user/items (e.g., text or images) based on pre-trained modality encoders and LLMs. Despite their effectiveness under cold scenarios, we observe that they underperform simple traditional collaborative filtering models under warm scenarios due to the lack of collaborative knowledge. In this work, we propose an efficient All-round LLM-based Recommender system, called A-LLMRec, that excels not only in the cold scenario but also in the warm scenario. Our main idea is to enable an LLM to directly leverage the collaborative knowledge contained in a pre-trained state-of-the-art CF-RecSys so that the emergent ability of the LLM as well as the high-quality user/item embeddings that are already trained by the state-of-the-art CF-RecSys can be jointly exploited. This approach yields two advantages: (1) model-agnostic, allowing for integration with various existing CF-RecSys, and (2) efficiency, eliminating the extensive fine-tuning typically required for LLM-based recommenders. Our extensive experiments on various real-world datasets demonstrate the superiority of A-LLMRec in various scenarios, including cold/warm, few-shot, cold user, and cross-domain scenarios. Beyond the recommendation task, we also show the potential of A-LLMRec in generating natural language outputs based on the understanding of the collaborative knowledge by performing a favorite genre prediction task. Our code is available at https://github.com/ghdtjr/A-LLMRec .

LLMs and Collaborative Filtering: Advancements in Recommender Systems

The paper "LLMs meet Collaborative Filtering: An Efficient All-round LLM-based Recommender System" by Sein Kim et al. explores the integration of LLMs with collaborative filtering approaches to build a robust recommender system, named A-LLMRec. This system addresses the limitations of traditional collaborative filtering recommender systems (CF-RecSys) in cold-start scenarios and proposes a novel architecture that excels in both cold and warm scenarios by leveraging collaborative knowledge and textual information.

Collaborative filtering recommender systems have been instrumental in improving user experiences across platforms, yet they face significant challenges in scenarios with sparse user-item interactions. Previous research efforts have focused on blending modality information—such as text or images—through pre-trained modality encoders like BERT and Vision-Transformer to combat these challenges, particularly under cold-start scenarios. However, these methods often underperform in warm scenarios when compared to traditional collaborative filtering models due to inadequate exploitation of collaborative knowledge derived from extensive user-item interactions.

To remedy this, the authors introduce A-LLMRec, which innovatively integrates a state-of-the-art CF-RecSys with an LLM, thereby exploiting the inherent abilities of LLMs alongside high-quality user/item embeddings from CF-RecSys. A distinguishing feature of A-LLMRec is its model-agnostic nature, which facilitates its integration with any existing recommender systems, thereby enhancing its practical applicability. Moreover, the system is designed for efficiency by eliminating the necessity for extensive fine-tuning of LLM-based recommenders.

The experiments conducted on various real-world datasets highlight the superior performance of A-LLMRec in multiple scenarios including cold/warm, few-shot, cold user, and cross-domain contexts. The paper reports robust results, displaying significant improvements over both traditional and contemporary models across these settings. Particularly, the ability of A-LLMRec to perform well with few interactions or in cross-domain recommendations underscores its versatility and robustness.

The implications of this research are multifaceted. Practically, A-LLMRec provides a scalable solution capable of handling various recommendation scenarios effectively, thus offering substantial improvements in user experience on platforms reliant on recommender systems. Theoretically, this work demonstrates the merits of symbiotically using collaborative filtering's latent knowledge with LLM's contextual understanding, enriching the domain literature by proposing an innovative way of enhancing traditional systems with modern deep learning architectures.

Future directions could explore deeper integrations of personalized natural language interactions and recommendations by leveraging advances in multi-modal LLMs or further optimizing the balance between collaborative knowledge and modality influences. Additionally, the implications of the system's model-agnostic properties could lead to wider adoption and experimentation on different recommendation tasks beyond those considered in this paper. Overall, A-LLMRec represents a significant step forward in enhancing recommender systems through a nuanced blend of collaborative and machine learning advancements.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (53)
  1. Controlling Popularity Bias in Learning-to-Rank Recommendation. In Proceedings of the Eleventh ACM Conference on Recommender Systems (Como, Italy) (RecSys ’17). Association for Computing Machinery, New York, NY, USA, 42–46. https://doi.org/10.1145/3109859.3109912
  2. Tallrec: An effective and efficient tuning framework to align large language model with recommendation. arXiv preprint arXiv:2305.00447 (2023).
  3. Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150 (2020).
  4. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
  5. A probabilistic model for using social networks in personalized item recommendation. In Proceedings of the 9th ACM Conference on Recommender Systems. 43–50.
  6. AutoDebias: Learning to Debias for Recommendation (SIGIR ’21). Association for Computing Machinery, New York, NY, USA, 21–30. https://doi.org/10.1145/3404835.3462919
  7. Where you like to go next: successive point-of-interest recommendation. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence (Beijing, China) (IJCAI ’13). AAAI Press, 2605–2611.
  8. Robert G. Cooper and Scott J. Edgett. 2012. Best Practices in the Idea-to-Launch Process and Its Governance. Research Technology Management 55, 2 (2012), 43–54. https://www.jstor.org/stable/26586220
  9. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, Jill Burstein, Christy Doran, and Thamar Solorio (Eds.). Association for Computational Linguistics, Minneapolis, Minnesota, 4171–4186. https://doi.org/10.18653/v1/N19-1423
  10. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations.
  11. Invariant Representation Learning for Multimedia Recommendation. In Proceedings of the 30th ACM International Conference on Multimedia (¡conf-loc¿, ¡city¿Lisboa¡/city¿, ¡country¿Portugal¡/country¿, ¡/conf-loc¿) (MM ’22). Association for Computing Machinery, New York, NY, USA, 619–628. https://doi.org/10.1145/3503161.3548405
  12. Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv preprint arXiv:2303.14524 (2023).
  13. Ruining He and Julian McAuley. 2016. Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering. In Proceedings of the 25th International Conference on World Wide Web (Montréal, Québec, Canada) (WWW ’16). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE, 507–517. https://doi.org/10.1145/2872427.2883037
  14. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval. 639–648.
  15. Neural collaborative filtering. In Proceedings of the 26th international conference on world wide web. 173–182.
  16. Large language models as zero-shot conversational recommenders. In Proceedings of the 32nd ACM international conference on information and knowledge management. 720–730.
  17. Session-based recommendations with recurrent neural networks. arXiv preprint arXiv:1511.06939 (2015).
  18. LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations.
  19. Collaborative filtering for implicit feedback datasets. In 2008 Eighth IEEE international conference on data mining. Ieee, 263–272.
  20. Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In 2018 IEEE international conference on data mining (ICDM). IEEE, 197–206.
  21. Task Relation-aware Continual User Representation Learning. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’23). Association for Computing Machinery, New York, NY, USA, 1107–1119. https://doi.org/10.1145/3580305.3599516
  22. Matrix factorization techniques for recommender systems. Computer 42, 8 (2009), 30–37.
  23. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems, F. Pereira, C.J. Burges, L. Bottou, and K.Q. Weinberger (Eds.), Vol. 25. Curran Associates, Inc.
  24. Text Is All You Need: Learning Language Representations for Sequential Recommendation (KDD ’23). Association for Computing Machinery, New York, NY, USA, 1258–1267. https://doi.org/10.1145/3580305.3599519
  25. CTRL: Connect Tabular and Language Model for CTR Prediction. arXiv preprint arXiv:2306.02841 (2023).
  26. Xiang Lisa Li and Percy Liang. 2021. Prefix-Tuning: Optimizing Continuous Prompts for Generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers).
  27. Noninvasive self-attention for side information fusion in sequential recommendation. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 35. 4249–4256.
  28. Disentangled multimodal representation learning for recommendation. IEEE Transactions on Multimedia (2022).
  29. Multi-Modal Contrastive Pre-training for Recommendation. In Proceedings of the 2022 International Conference on Multimedia Retrieval (Newark, NJ, USA) (ICMR ’22). Association for Computing Machinery, New York, NY, USA, 99–108. https://doi.org/10.1145/3512527.3531378
  30. Chih-Chao Ma. 2008. A guide to singular value decomposition for collaborative filtering. Computer (Long Beach, CA) 2008 (2008), 1–14.
  31. Image-Based Recommendations on Styles and Substitutes (SIGIR ’15). Association for Computing Machinery, New York, NY, USA, 43–52. https://doi.org/10.1145/2766462.2767755
  32. Image-based recommendations on styles and substitutes. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval. 43–52.
  33. Andriy Mnih and Russ R Salakhutdinov. 2007. Probabilistic matrix factorization. Advances in neural information processing systems 20 (2007).
  34. MUSE: Music Recommender System with Shuffle Play Recommendation Enhancement. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM ’23). Association for Computing Machinery, New York, NY, USA, 1928–1938. https://doi.org/10.1145/3583780.3614976
  35. Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084 (2019).
  36. Factorizing personalized Markov chains for next-basket recommendation. In Proceedings of the 19th International Conference on World Wide Web (Raleigh, North Carolina, USA) (WWW ’10). Association for Computing Machinery, New York, NY, USA, 811–820. https://doi.org/10.1145/1772690.1772773
  37. Large language models are competitive near cold-start recommenders for language-and item-based preferences. In Proceedings of the 17th ACM conference on recommender systems. 890–896.
  38. Item-based collaborative filtering recommendation algorithms. In Proceedings of the 10th international conference on World Wide Web. 285–295.
  39. Autorec: Autoencoders meet collaborative filtering. In Proceedings of the 24th international conference on World Wide Web. 111–112.
  40. BERT4Rec: Sequential recommendation with bidirectional encoder representations from transformer. In Proceedings of the 28th ACM international conference on information and knowledge management. 1441–1450.
  41. Jiaxi Tang and Ke Wang. 2018. Personalized top-n sequential recommendation via convolutional sequence embedding. In Proceedings of the eleventh ACM international conference on web search and data mining. 565–573.
  42. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023).
  43. DropoutNet: Addressing Cold Start in Recommender Systems. In Advances in Neural Information Processing Systems, I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.), Vol. 30. Curran Associates, Inc. https://proceedings.neurips.cc/paper_files/paper/2017/file/dbd22ba3bd0df8f385bdac3e9f8be207-Paper.pdf
  44. Lei Wang and Ee-Peng Lim. 2023. Zero-Shot Next-Item Recommendation using Large Pretrained Language Models. arXiv preprint arXiv:2304.03153 (2023).
  45. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022).
  46. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022), 24824–24837.
  47. MMGCN: Multi-modal Graph Convolution Network for Personalized Recommendation of Micro-video (MM ’19). Association for Computing Machinery, New York, NY, USA, 1437–1445. https://doi.org/10.1145/3343031.3351034
  48. A Survey on Large Language Models for Recommendation. arXiv preprint arXiv:2305.19860 (2023).
  49. DCBT: A Simple But Effective Way for Unified Warm and Cold Recommendation. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’23). Association for Computing Machinery, New York, NY, USA, 3369–3373. https://doi.org/10.1145/3539618.3591856
  50. A simple convolutional generative network for next item recommendation. In Proceedings of the twelfth ACM international conference on web search and data mining. 582–590.
  51. Where to Go Next for Recommender Systems? ID- vs. Modality-based Recommender Models Revisited (SIGIR ’23). Association for Computing Machinery, New York, NY, USA, 2639–2649. https://doi.org/10.1145/3539618.3591932
  52. Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022).
  53. SVD-based incremental approaches for recommender systems. J. Comput. System Sci. 81, 4 (2015), 717–733.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Sein Kim (11 papers)
  2. Hongseok Kang (5 papers)
  3. Seungyoon Choi (4 papers)
  4. Donghyun Kim (129 papers)
  5. Minchul Yang (3 papers)
  6. Chanyoung Park (83 papers)
Citations (4)
X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com