Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

How Can Recommender Systems Benefit from Large Language Models: A Survey (2306.05817v6)

Published 9 Jun 2023 in cs.IR and cs.AI
How Can Recommender Systems Benefit from Large Language Models: A Survey

Abstract: With the rapid development of online services, recommender systems (RS) have become increasingly indispensable for mitigating information overload. Despite remarkable progress, conventional recommendation models (CRM) still have some limitations, e.g., lacking open-world knowledge, and difficulties in comprehending users' underlying preferences and motivations. Meanwhile, LLMs (LLM) have shown impressive general intelligence and human-like capabilities, which mainly stem from their extensive open-world knowledge, reasoning ability, as well as their comprehension of human culture and society. Consequently, the emergence of LLM is inspiring the design of recommender systems and pointing out a promising research direction, i.e., whether we can incorporate LLM and benefit from their knowledge and capabilities to compensate for the limitations of CRM. In this paper, we conduct a comprehensive survey on this research direction from the perspective of the whole pipeline in real-world recommender systems. Specifically, we summarize existing works from two orthogonal aspects: where and how to adapt LLM to RS. For the WHERE question, we discuss the roles that LLM could play in different stages of the recommendation pipeline, i.e., feature engineering, feature encoder, scoring/ranking function, user interaction, and pipeline controller. For the HOW question, we investigate the training and inference strategies, resulting in two fine-grained taxonomy criteria, i.e., whether to tune LLM or not, and whether to involve conventional recommendation models for inference. Then, we highlight key challenges in adapting LLM to RS from three aspects, i.e., efficiency, effectiveness, and ethics. Finally, we summarize the survey and discuss the future prospects. We actively maintain a GitHub repository for papers and other related resources: https://github.com/CHIANGEL/Awesome-LLM-for-RecSys/.

An Expert Analysis of "How Can Recommender Systems Benefit from LLMs: A Survey"

This survey paper conducted by Lin et al. from Shanghai Jiao Tong University and Noah's Ark Lab, Huawei, offers a comprehensive examination of the integration and potential advantages of incorporating LLMs into recommender systems (RS). The paper carries out an extensive literature review, identifying the fundamental ways LLMs can augment various stages of the RS pipeline and addressing the core challenges encountered during this integration.

LLMs in Recommender Systems: Where and How

The authors dissect the recommender system pipeline into distinct stages: feature engineering, feature encoding, scoring/ranking function, user interaction, and pipeline control. At each stage, the integration of LLMs offers unique potential benefits.

  • Feature Engineering and Encoding: LLMs can augment feature engineering by generating rich, auxiliary textual features, thereby enhancing user-item profiles. This addresses the data sparsity and domain knowledge gaps inherent in traditional models.
  • Scoring/Ranking: Leveraging LLM capabilities in logical and commonsense reasoning improves scoring mechanisms, enabling more effective personalization through sophisticated user interaction models.
  • User Interaction: LLMs enable dynamic, conversational interaction modes that adapt to user inputs in real-time, offering a more natural interface.
  • Pipeline Control: The enhanced control mechanisms provided by LLMs enable an adaptable, refined control of recommendation processes through advanced reasoning and decision-making capabilities.

The survey employs a nuanced classification framework based on two orthogonal dimensions: whether to fine-tune LLMs during model training and whether conventional recommendation models are engaged during inference. This structure allows researchers to understand how LLMs can best fit into existing systems and the trade-offs involved.

Implications and Challenges

The integration of LLMs with recommender systems promises substantial improvements in both personalized user experience and overall model performance. By incorporating LLMs, recommender systems can mitigate existing challenges such as information overload and limited domain adaptation, thereby enhancing both efficiency and effectiveness.

However, the paper also outlines several challenges faced when adapting LLMs to RS contexts:

  • Efficiency: The computational overhead associated with training and deploying LLMs is significant, necessitating thoughtful strategies such as parameter-efficient finetuning and pre-computing techniques to ensure practical feasibility.
  • Effectiveness: While LLMs offer enhanced representation capabilities, effectively modeling long, contextual sequences and integrating ID features remain open questions that require innovative solutions.
  • Ethics and Fairness: Adopting LLMs raises ethical issues, including biases present in training corpuses and decision transparency, demanding robust fairness measures and explainability solutions.

Conclusion and Future Directions

The paper successfully identifies the convergence between LLMs and RS as a promising domain of research. The methodological insights and detailed analysis provide clear pathways for future developments, encouraging exploration into custom LLM architectures tailored to specific recommendation contexts and the development of benchmarks to systematically measure integration outcomes.

In conclusion, the work by Lin et al. offers a pivotal reference point for researchers and practitioners aiming to harness LLMs within recommender systems, pointing toward a future where these systems not only meet but surpass current capability thresholds. Future endeavors in this domain should concentrate on balancing LLM strengths with the requisite computational feasibility while ensuring ethical usage, setting the foundation for more adaptive, intelligent, and user-centric recommendation applications.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (87)
  1. Tallrec: An effective and efficient tuning framework to align large language model with recommendation. arXiv preprint arXiv:2305.00447, 2023.
  2. Language models are realistic tabular data generators. In The Eleventh International Conference on Learning Representations, 2023.
  3. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  4. Privacy-preserving recommender systems with synthetic query generation using differentially private large language models. arXiv preprint arXiv:2305.05973, 2023.
  5. The lottery ticket hypothesis for pre-trained bert networks. Advances in neural information processing systems, 33:15834–15846, 2020.
  6. Knowledge graph completion models are few-shot learners: An empirical study of relation labeling in e-commerce with llms. arXiv preprint arXiv:2305.09858, 2023.
  7. Zheng Chen. Palr: Personalization aware llms for recommendation. arXiv preprint arXiv:2305.07622, 2023.
  8. Large language models for user interest journeys. arXiv preprint arXiv:2305.15498, 2023.
  9. M6-rec: Generative pretrained language models are open-ended recommender systems. arXiv preprint arXiv:2205.08084, 2022.
  10. An adversarial imitation click model for information retrieval. In Proceedings of the Web Conference 2021, pages 1809–1820, 2021.
  11. Uncovering chatgpt’s capabilities in recommender systems. arXiv preprint arXiv:2305.02182, 2023.
  12. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  13. Cogltx: Applying bert to long texts. Advances in Neural Information Processing Systems, 33:12792–12804, 2020.
  14. Zero-shot recommender systems. arXiv preprint arXiv:2105.08318, 2021.
  15. Leveraging large language models in conversational recommender systems. arXiv preprint arXiv:2305.07961, 2023.
  16. Exploring adapter-based transfer learning for recommender systems: Empirical studies and practical insights. arXiv preprint arXiv:2305.15036, 2023.
  17. An f-shape click model for information retrieval on multi-block mobile pages. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, pages 1057–1065, 2023.
  18. Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv preprint arXiv:2303.14524, 2023.
  19. Recommendation as language processing (rlp): A unified pretrain, personalized prompt & predict paradigm (p5). In Proceedings of the 16th ACM Conference on Recommender Systems, pages 299–315, 2022.
  20. Vip5: Towards multimodal foundation models for recommendation. arXiv preprint arXiv:2305.14302, 2023.
  21. Deepfm: a factorization-machine based neural network for ctr prediction. arXiv preprint arXiv:1703.04247, 2017.
  22. Ptm4tag: sharpening tag recommendation of stack overflow posts with pre-trained models. In Proceedings of the 30th IEEE/ACM International Conference on Program Comprehension, pages 1–11, 2022.
  23. Tabllm: Few-shot classification of tabular data with large language models. In International Conference on Artificial Intelligence and Statistics, pages 5549–5581. PMLR, 2023.
  24. Towards universal sequence representation learning for recommender systems. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 585–593, 2022.
  25. Learning vector-quantized item representation for transferable sequential recommenders. In Proceedings of the ACM Web Conference 2023, pages 1162–1171, 2023.
  26. Large language models are zero-shot rankers for recommender systems. arXiv preprint arXiv:2305.08845, 2023.
  27. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685, 2021.
  28. Up5: Unbiased foundation model for fairness-aware recommendation. arXiv preprint arXiv:2305.12090, 2023.
  29. How to index item ids for recommendation foundation models. arXiv preprint arXiv:2305.06569, 2023.
  30. Jie Huang and Kevin Chen-Chuan Chang. Towards reasoning in large language models: A survey. arXiv preprint arXiv:2212.10403, 2022.
  31. Tinybert: Distilling bert for natural language understanding. arXiv preprint arXiv:1909.10351, 2019.
  32. Do llms understand user preferences? evaluating llms on user rating prediction. arXiv preprint arXiv:2305.06474, 2023.
  33. Taggpt: Large language models are zero-shot multimodal taggers. arXiv preprint arXiv:2304.03022, 2023.
  34. Text is all you need: Learning language representations for sequential recommendation. arXiv preprint arXiv:2305.13731, 2023.
  35. Gpt4rec: A generative framework for personalized recommendation and user interests interpretation. arXiv preprint arXiv:2304.03879, 2023.
  36. Exploring the upper limits of text-based collaborative filtering using large language models: Discoveries and insights. arXiv preprint arXiv:2305.11700, 2023.
  37. Ctrl: Connect tabular and language model for ctr prediction. arXiv preprint arXiv:2306.02841, 2023.
  38. Pbnr: Prompt-based news recommender system. arXiv preprint arXiv:2304.07862, 2023.
  39. A preliminary study of chatgpt on news recommendation: Personalization, provider fairness, fake news. arXiv preprint arXiv:2306.10702, 2023.
  40. Sparks of artificial general recommender (agr): Early experiments with chatgpt. arXiv preprint arXiv:2305.04518, 2023.
  41. A graph-enhanced click model for web search. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 1259–1268, 2021.
  42. Personalized fairness-aware re-ranking for microlending. In Proceedings of the 13th ACM conference on recommender systems, pages 467–471, 2019.
  43. Pre-trained language model for web-scale retrieval in baidu search. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pages 3365–3375, 2021.
  44. Ptab: Using the pre-trained language model for modeling tabular data. arXiv preprint arXiv:2209.08060, 2022.
  45. Boosting deep ctr prediction with a plug-and-play pre-trainer for news recommendation. In Proceedings of the 29th International Conference on Computational Linguistics, pages 2823–2833, 2022.
  46. Is chatgpt a good recommender? a preliminary study. arXiv preprint arXiv:2304.10149, 2023.
  47. Pre-train, prompt and recommendation: A comprehensive survey of language modelling paradigm adaptations in recommender systems. arXiv preprint arXiv:2302.03735, 2023.
  48. A first look at llm-powered generative news recommendation. arXiv preprint arXiv:2305.06566, 2023.
  49. Large language model is not a good few-shot information extractor, but a good reranker for hard samples! arXiv preprint arXiv:2303.08559, 2023.
  50. Unitrec: A unified text-to-text transformer and joint contrastive learning framework for text-based recommendation. arXiv preprint arXiv:2305.15756, 2023.
  51. Ctr-bert: Cost-effective knowledge distillation for billion-parameter teacher models. In NeurIPS Efficient Natural Language and Speech Processing Workshop, 2021.
  52. Large language model augmented narrative driven recommendations. arXiv preprint arXiv:2306.02250, 2023.
  53. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
  54. Fairrec: Two-sided fairness for personalized recommendations in two-sided platforms. In Proceedings of the web conference 2020, pages 1194–1204, 2020.
  55. Generative sequential recommendation with gptrec. arXiv preprint arXiv:2306.11114, 2023.
  56. U-bert: Pre-training user representations for improved recommendation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 4320–4327, 2021.
  57. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  58. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485–5551, 2020.
  59. Improving code example recommendations on informal documentation using bert and query-aware lsh: A comparative study. arXiv preprint arXiv:2305.03017, 2023.
  60. Zero-shot recommendation as language modeling. In Advances in Information Retrieval: 44th European Conference on IR Research, ECIR 2022, Stavanger, Norway, April 10–14, 2022, Proceedings, Part II, pages 223–230. Springer, 2022.
  61. Fairness of exposure in rankings. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 2219–2228, 2018.
  62. Bert4rec: Sequential recommendation with bidirectional encoder representations from transformer. In Proceedings of the 28th ACM international conference on information and knowledge management, pages 1441–1450, 2019.
  63. Is chatgpt good at search? investigating large language models as re-ranking agent. arXiv preprint arXiv:2304.09542, 2023.
  64. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  65. Zero-shot next-item recommendation using large pretrained language models. arXiv preprint arXiv:2304.03153, 2023.
  66. Multi-passage bert: A globally normalized bert model for open-domain question answering. arXiv preprint arXiv:1908.08167, 2019.
  67. Transrec: Learning transferable recommendation from mixture-of-modality feedback. arXiv preprint arXiv:2206.06190, 2022.
  68. Anypredict: Foundation model for tabular prediction. arXiv preprint arXiv:2305.12081, 2023.
  69. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682, 2022.
  70. Empowering news recommendation with pre-trained language models. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 1652–1656, 2021.
  71. Mm-rec: Visiolinguistic model empowered multimodal news recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2560–2564, 2022.
  72. A bird’s-eye view of reranking: from list level to page level. In Proceedings of the Sixteenth ACM International Conference on Web Search and Data Mining, pages 1075–1083, 2023.
  73. Towards open-world recommendation with knowledge augmentation from large language models. arXiv preprint arXiv:2306.10933, 2023.
  74. Self-supervised learning for recommender systems: A survey. arXiv preprint arXiv:2203.15876, 2022.
  75. Tiny-newsrec: Effective and efficient plm-based news recommendation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 5478–5489, 2022.
  76. Where to go next for recommender systems? id-vs. modality-based recommender models revisited. arXiv preprint arXiv:2303.13835, 2023.
  77. Q8bert: Quantized 8bit bert. In 2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing-NeurIPS Edition (EMC2-NIPS), pages 36–39. IEEE, 2019.
  78. Prompt learning for news recommendation. arXiv preprint arXiv:2304.05263, 2023.
  79. Unbert: User-news matching bert for news recommendation. In IJCAI, pages 3356–3362, 2021.
  80. Language models as recommender systems: Evaluations and limitations. 2021.
  81. Twhin-bert: A socially-enriched pre-trained language model for multilingual tweet representations. arXiv preprint arXiv:2209.07562, 2022.
  82. Is chatgpt fair for recommendation? evaluating fairness in large language model recommendation. arXiv preprint arXiv:2305.07609, 2023.
  83. Recommendation as instruction following: A large language model empowered recommendation approach. arXiv preprint arXiv:2305.07001, 2023.
  84. A survey of large language models. arXiv preprint arXiv:2303.18223, 2023.
  85. Bookgpt: A general framework for book recommendation empowered by large language model. arXiv preprint arXiv:2305.15673, 2023.
  86. Lima: Less is more for alignment. arXiv preprint arXiv:2305.11206, 2023.
  87. Pre-trained language model based ranking in baidu search. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pages 4014–4022, 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (14)
  1. Jianghao Lin (47 papers)
  2. Xinyi Dai (32 papers)
  3. Yunjia Xi (21 papers)
  4. Weiwen Liu (59 papers)
  5. Bo Chen (309 papers)
  6. Xiangyang Li (58 papers)
  7. Chenxu Zhu (14 papers)
  8. Huifeng Guo (60 papers)
  9. Yong Yu (219 papers)
  10. Ruiming Tang (171 papers)
  11. Weinan Zhang (322 papers)
  12. Hao Zhang (947 papers)
  13. Yong Liu (721 papers)
  14. Chuhan Wu (86 papers)
Citations (150)
X Twitter Logo Streamline Icon: https://streamlinehq.com