Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Large Language Models Enhanced Collaborative Filtering (2403.17688v2)

Published 26 Mar 2024 in cs.IR
Large Language Models Enhanced Collaborative Filtering

Abstract: Recent advancements in LLMs have attracted considerable interest among researchers to leverage these models to enhance Recommender Systems (RSs). Existing work predominantly utilizes LLMs to generate knowledge-rich texts or utilizes LLM-derived embeddings as features to improve RSs. Although the extensive world knowledge embedded in LLMs generally benefits RSs, the application can only take limited number of users and items as inputs, without adequately exploiting collaborative filtering information. Considering its crucial role in RSs, one key challenge in enhancing RSs with LLMs lies in providing better collaborative filtering information through LLMs. In this paper, drawing inspiration from the in-context learning and chain of thought reasoning in LLMs, we propose the LLMs enhanced Collaborative Filtering (LLM-CF) framework, which distils the world knowledge and reasoning capabilities of LLMs into collaborative filtering. We also explored a concise and efficient instruction-tuning method, which improves the recommendation capabilities of LLMs while preserving their general functionalities (e.g., not decreasing on the LLM benchmark). Comprehensive experiments on three real-world datasets demonstrate that LLM-CF significantly enhances several backbone recommendation models and consistently outperforms competitive baselines, showcasing its effectiveness in distilling the world knowledge and reasoning capabilities of LLM into collaborative filtering.

Enhancing Recommender Systems with LLM Collaborative Filtering

Introduction to LLM-CF Framework

The recommender systems (RSs) domain has seen significant advancements with the integration of LLMs, particularly in generating knowledge-rich texts and embeddings to improve recommendation quality. However, a gap exists in fully leveraging LLMs for collaborative filtering (CF) due to their limited capacity to process extensive user and item data in a single prompt. Addressing this challenge, the proposed LLMs enhanced Collaborative Filtering (LLM-CF) framework introduces a novel approach that distils LLMs' world knowledge and reasoning capabilities directly into CF. LLM-CF operates by fine-tuning an LLM with a concise instruction-tuning method that balances the model's recommendation-specific and general functionalities. Through comprehensive experiments on three real-world datasets, LLM-CF is demonstrated to significantly enhance various backbone recommendation models and consistently outperform competitive baselines.

Key Contributions and Framework Overview

The LLM-CF framework comprises two main parts: an offline service for fine-tuning and generating CoT (Chain of Thought) reasoning and an online service that incorporates the distilled LLM knowledge into RSs. The framework's development was motivated by the limitations of existing LLM applications in RSs, which either underutilize the collaborative filtering information or suffer from deployment inefficiencies due to LLMs' extensive parameters.

  • Distillation of World Knowledge: LLM-CF fine-tunes an LLM on a mix of recommendation and general instruction-tuning data to obtain a model, RecGen-LLaMA, which maintains an optimal balance between general functionalities and recommendation capabilities. This model is then utilized to generate CoT reasoning for a subset of training examples, creating an in-context CoT dataset.
  • Learning Collaborative Filtering Features: The in-context CoT examples are retrieved based on similarity with current recommendation features, and the In-context Chain of Thought (ICT) module processes these examples to extract world-knowledge and reasoning guided collaborative filtering feature. This process leverages transformer decoder layers for the in-context learning of CoT examples, encapsulating the enhanced CF feature into existing RSs efficiently.

Experimental Validation and Implications

The framework underwent extensive testing on three public datasets, demonstrating its ability to significantly improve several conventional recommendation models across ranking and retrieval tasks. The results underscore LLM-CF's efficacy in leveraging LLMs for distilling world knowledge and reasoning into RSs, reflecting on several crucial insights:

  • Enhancement of Collaborative Filtering: The integration of CoT reasoning and world knowledge from LLMs directly into collaborative filtering processes leads to a substantial improvement in recommendation accuracy.
  • Deployment Efficiency: By decoupling LLM generation from the RS's online services through an efficient, two-part framework, LLM-CF addresses a significant challenge in deploying LLM-enhanced RSs, offering a scalable solution.
  • Model Versatility: The application of LLM-CF across various backbone models and tasks (both ranking and retrieval) highlights its versatility and potential to generalize across different recommendation scenarios.

Future Directions

The demonstrated success of the LLM-CF framework opens several avenues for further research:

  • Exploration of LLM Capacities: Future studies could explore the integration of other LLM capabilities, such as sentiment analysis or question-answering, into the recommendation process.
  • Scalability and Efficiency: Investigating methods to further optimize the efficiency of LLM-CF, particularly in processing larger datasets or real-time recommendation scenarios, remains a critical area.
  • Cross-Domain Applications: Extending the framework's application to other domains beyond traditional RSs, such as content curation or personalized search, could further validate its utility and impact.

In summary, the LLM-CF framework represents a significant step forward in utilizing LLMs to enhance RSs, particularly through the effective distillation of collaborative filtering information. Its implications for both practical deployments and theoretical advancements in AI and RSs highlight the potential of merging LLMs' cognitive capabilities with domain-specific applications.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (56)
  1. Better Fine-Tuning by Reducing Representational Collapse. ArXiv abs/2008.03156 (2020).
  2. TALLRec: An Effective and Efficient Tuning Framework to Align Large Language Model with Recommendation. Proceedings of the 17th ACM Conference on Recommender Systems (2023).
  3. TWIN: TWo-stage Interest Network for Lifelong User Behavior Modeling in CTR Prediction at Kuaishou. Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (2023).
  4. A survey of chain of thought reasoning: Advances, frontiers and future. arXiv preprint arXiv:2309.15402 (2023).
  5. Leveraging Large Language Models for Pre-trained Recommender Systems. arXiv:2308.10837 [cs.IR]
  6. Deep Neural Networks for YouTube Recommendations. Proceedings of the 10th ACM Conference on Recommender Systems (2016).
  7. Why Can GPT Learn In-Context? Language Models Secretly Perform Gradient Descent as Meta-Optimizers. In Findings of the Association for Computational Linguistics: ACL 2023. Association for Computational Linguistics, Toronto, Canada, 4005–4019.
  8. Uncovering ChatGPT’s Capabilities in Recommender Systems. Proceedings of the 17th ACM Conference on Recommender Systems (2023). https://api.semanticscholar.org/CorpusID:258461170
  9. How Abilities in Large Language Models are Affected by Supervised Fine-tuning Data Composition. ArXiv abs/2310.05492 (2023).
  10. A survey for in-context learning. arXiv preprint arXiv:2301.00234 (2022).
  11. Recommendation as Language Processing (RLP): A Unified Pretrain, Personalized Prompt & Predict Paradigm (P5). In Proceedings of the 16th ACM Conference on Recommender Systems (Seattle, WA, USA) (RecSys ’22). Association for Computing Machinery, New York, NY, USA, 299–315. https://doi.org/10.1145/3523227.3546767
  12. DeepFM: A Factorization-Machine Based Neural Network for CTR Prediction. In Proceedings of the 26th International Joint Conference on Artificial Intelligence (Melbourne, Australia) (IJCAI’17). AAAI Press, 1725–1731.
  13. Leveraging Large Language Models for Sequential Recommendation. Proceedings of the 17th ACM Conference on Recommender Systems (2023). https://api.semanticscholar.org/CorpusID:261823711
  14. Ruining He and Julian McAuley. 2016. Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering. In Proceedings of the 25th International Conference on World Wide Web (Montréal, Québec, Canada) (WWW ’16). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE, 507–517. https://doi.org/10.1145/2872427.2883037
  15. Measuring massive multitask language understanding. arXiv preprint arXiv:2009.03300 (2020).
  16. Balázs Hidasi and Alexandros Karatzoglou. 2018. Recurrent Neural Networks with Top-k Gains for Session-based Recommendations. In CIKM. ACM, 843–852.
  17. Distilling the Knowledge in a Neural Network. arXiv:1503.02531 [stat.ML]
  18. LoRA: Low-Rank Adaptation of Large Language Models. In ICLR. OpenReview.net.
  19. NEFTune: Noisy Embeddings Improve Instruction Finetuning. arXiv preprint arXiv:2310.05914 (2023).
  20. Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In 2018 IEEE international conference on data mining (ICDM). IEEE, 197–206.
  21. Do LLMs Understand User Preferences? Evaluating LLMs On User Rating Prediction. ArXiv abs/2305.06474 (2023). https://api.semanticscholar.org/CorpusID:258615591
  22. Reformer: The Efficient Transformer. In International Conference on Learning Representations.
  23. XDeepFM: Combining Explicit and Implicit Feature Interactions for Recommender Systems. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (London, United Kingdom) (KDD ’18). Association for Computing Machinery, New York, NY, USA, 1754–1763. https://doi.org/10.1145/3219819.3220023
  24. ReLLa: Retrieval-enhanced Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation. arXiv preprint arXiv:2308.11131 (2023).
  25. A First Look at LLM-Powered Generative News Recommendation. arXiv preprint arXiv:2305.06566 (2023).
  26. Kai Mei and Yongfeng Zhang. 2023. LightLM: A Lightweight Deep and Narrow Language Model for Generative Recommendation. arXiv:2310.17488 (2023).
  27. Marius Muja and David G. Lowe. 2009. Fast Approximate Nearest Neighbors with Automatic Algorithm Configuration. In International Conference on Computer Vision Theory and Applications.
  28. Large Language Model Augmented Narrative Driven Recommendations. Proceedings of the 17th ACM Conference on Recommender Systems (2023).
  29. OpenAI. 2023. GPT-4 Technical Report. ArXiv abs/2303.08774 (2023). https://api.semanticscholar.org/CorpusID:257532815
  30. Instruction tuning with gpt-4. arXiv preprint arXiv:2304.03277 (2023).
  31. Search-based User Interest Modeling with Lifelong Sequential Behavior Data for Click-Through Rate Prediction. Proceedings of the 29th ACM International Conference on Information & Knowledge Management (2020).
  32. Retrieval & Interaction Machine for Tabular Data Prediction. Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining (2021).
  33. Zero: Memory optimizations toward training trillion parameter models. In SC20: International Conference for High Performance Computing, Networking, Storage and Analysis. IEEE, 1–16.
  34. HM-ANN: Efficient Billion-Point Nearest Neighbor Search on Heterogeneous Memory. In Proceedings of the 34th International Conference on Neural Information Processing Systems (Vancouver, BC, Canada) (NIPS’20). Curran Associates Inc., Red Hook, NY, USA, Article 895, 13 pages.
  35. Ruifeng Ren and Yong Liu. 2023. In-context Learning with Transformer Is Really Equivalent to a Contrastive Learning Pattern. arXiv preprint arXiv:2310.13220 (2023).
  36. Large language models are competitive near cold-start recommenders for language-and item-based preferences. In Proceedings of the 17th ACM conference on recommender systems. 890–896.
  37. AutoInt: Automatic Feature Interaction Learning via Self-Attentive Neural Networks. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management (Beijing, China) (CIKM ’19). Association for Computing Machinery, New York, NY, USA, 1161–1170. https://doi.org/10.1145/3357384.3357925
  38. Gemini: A family of highly capable multimodal models. arXiv preprint arXiv:2312.11805 (2023).
  39. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023).
  40. Deep & Cross Network for Ad Click Predictions. In Proceedings of the ADKDD’17 (Halifax, NS, Canada) (ADKDD’17). Association for Computing Machinery, New York, NY, USA, Article 12, 7 pages. https://doi.org/10.1145/3124749.3124754
  41. DCN V2: Improved Deep & Cross Network and Practical Lessons for Web-Scale Learning to Rank Systems. In Proceedings of the Web Conference 2021 (Ljubljana, Slovenia) (WWW ’21). Association for Computing Machinery, New York, NY, USA, 1785–1797. https://doi.org/10.1145/3442381.3450078
  42. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022).
  43. LLMRec: Large Language Models with Graph Augmentation for Recommendation. ArXiv (2023).
  44. Robust Fine-Tuning of Zero-Shot Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 7959–7971.
  45. Session-Based Recommendation with Graph Neural Networks. Proceedings of the AAAI Conference on Artificial Intelligence 33, 01 (July 2019), 346–353. https://doi.org/10.1609/aaai.v33i01.3301346
  46. Towards Open-World Recommendation with Knowledge Augmentation from Large Language Models. arXiv preprint arXiv:2306.10933 (2023).
  47. C-Pack: Packaged Resources To Advance General Chinese Embedding. arXiv:2309.07597 [cs.CL]
  48. Heterogeneous Knowledge Fusion: A Novel Approach for Personalized Recommendation via LLM. Proceedings of the 17th ACM Conference on Recommender Systems (2023). https://api.semanticscholar.org/CorpusID:260682982
  49. HyPe: Better Pre-trained Language Model Fine-tuning with Hidden Representation Perturbation. In Annual Meeting of the Association for Computational Linguistics.
  50. LlamaRec: Two-Stage Recommendation using Large Language Models for Ranking. ArXiv abs/2311.02089 (2023). https://api.semanticscholar.org/CorpusID:265033634
  51. Generative job recommendations with large language model. arXiv preprint arXiv:2307.02157 (2023).
  52. Lima: Less is more for alignment. arXiv preprint arXiv:2305.11206 (2023).
  53. Deep Interest Network for Click-Through Rate Prediction. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (London, United Kingdom) (KDD ’18). Association for Computing Machinery, New York, NY, USA, 1059–1068. https://doi.org/10.1145/3219819.3219823
  54. S3-Rec: Self-Supervised Learning for Sequential Recommendation with Mutual Information Maximization. In CIKM ’20: The 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, October 19-23, 2020. ACM, 1893–1902.
  55. Filter-Enhanced MLP is All You Need for Sequential Recommendation. In Proceedings of the ACM Web Conference 2022 (Virtual Event, Lyon, France) (WWW ’22). Association for Computing Machinery, New York, NY, USA, 2388–2399. https://doi.org/10.1145/3485447.3512111
  56. Ensembled CTR Prediction via Knowledge Distillation. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management (Virtual Event, Ireland) (CIKM ’20). Association for Computing Machinery, New York, NY, USA, 2941–2958. https://doi.org/10.1145/3340531.3412704
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhongxiang Sun (21 papers)
  2. Zihua Si (12 papers)
  3. Xiaoxue Zang (28 papers)
  4. Kai Zheng (134 papers)
  5. Yang Song (298 papers)
  6. Xiao Zhang (435 papers)
  7. Jun Xu (397 papers)
Citations (5)
X Twitter Logo Streamline Icon: https://streamlinehq.com