Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Large Language Model Enhanced Conversational Recommender System (2308.06212v1)

Published 11 Aug 2023 in cs.IR and cs.CL

Abstract: Conversational recommender systems (CRSs) aim to recommend high-quality items to users through a dialogue interface. It usually contains multiple sub-tasks, such as user preference elicitation, recommendation, explanation, and item information search. To develop effective CRSs, there are some challenges: 1) how to properly manage sub-tasks; 2) how to effectively solve different sub-tasks; and 3) how to correctly generate responses that interact with users. Recently, LLMs have exhibited an unprecedented ability to reason and generate, presenting a new opportunity to develop more powerful CRSs. In this work, we propose a new LLM-based CRS, referred to as LLMCRS, to address the above challenges. For sub-task management, we leverage the reasoning ability of LLM to effectively manage sub-task. For sub-task solving, we collaborate LLM with expert models of different sub-tasks to achieve the enhanced performance. For response generation, we utilize the generation ability of LLM as a language interface to better interact with users. Specifically, LLMCRS divides the workflow into four stages: sub-task detection, model matching, sub-task execution, and response generation. LLMCRS also designs schema-based instruction, demonstration-based instruction, dynamic sub-task and model matching, and summary-based generation to instruct LLM to generate desired results in the workflow. Finally, to adapt LLM to conversational recommendations, we also propose to fine-tune LLM with reinforcement learning from CRSs performance feedback, referred to as RLPF. Experimental results on benchmark datasets show that LLMCRS with RLPF outperforms the existing methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (47)
  1. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
  2. A survey on dialogue systems: Recent advances and new frontiers. Acm Sigkdd Explorations Newsletter 19, 2 (2017), 25–35.
  3. Towards Knowledge-Based Recommender Dialog System. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. 1803–1813.
  4. Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311 (2022).
  5. Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416 (2022).
  6. Uncovering ChatGPT’s Capabilities in Recommender Systems. arXiv preprint arXiv:2305.02182 (2023).
  7. Advances and challenges in conversational recommender systems: A survey. AI Open 2 (2021), 100–126.
  8. Neural approaches to conversational AI. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 1371–1374.
  9. Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv preprint arXiv:2303.14524 (2023).
  10. Kalervo Järvelin and Jaana Kekäläinen. 2017. IR evaluation methods for retrieving highly relevant documents. In ACM SIGIR Forum, Vol. 51. 243–250.
  11. Recommendation as a Communication Game: Self-Supervised Bot-Play for Goal-oriented Dialogue. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. 1951–1961.
  12. Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of NAACL-HLT. 4171–4186.
  13. Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
  14. Estimation-action-reflection: Towards deep interaction between conversational and recommender systems. In Proceedings of the 13th International Conference on Web Search and Data Mining. 304–312.
  15. A Diversity-Promoting Objective Function for Neural Conversation Models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 110–119.
  16. Towards deep conversational recommendations. Advances in neural information processing systems 31 (2018).
  17. Yuxi Li. 2017. Deep reinforcement learning: An overview. arXiv preprint arXiv:1701.07274 (2017).
  18. Pre-train, prompt and recommendation: A comprehensive survey of language modelling paradigm adaptations in recommender systems. arXiv preprint arXiv:2302.03735 (2023).
  19. The flan collection: Designing data and methods for effective instruction tuning. arXiv preprint arXiv:2301.13688 (2023).
  20. ParlAI: A Dialog Research Software Platform. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. 79–84.
  21. Key-Value Memory Networks for Directly Reading Documents. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 1400–1409.
  22. Top-k learning to rank: labeling, ranking and evaluation. In Proceedings of SIGIR. 751–760.
  23. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems 35 (2022), 27730–27744.
  24. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics. 311–318.
  25. Language models are unsupervised multitask learners. (2018).
  26. Toolformer: Language models can teach themselves to use tools. arXiv preprint arXiv:2302.04761 (2023).
  27. Multi-grained Hypergraph Interest Modeling for Conversational Recommendation. arXiv preprint arXiv:2305.04798 (2023).
  28. Hugginggpt: Solving ai tasks with chatgpt and its friends in huggingface. arXiv preprint arXiv:2303.17580 (2023).
  29. Richard S Sutton and Andrew G Barto. 2018. Reinforcement learning: An introduction. MIT press.
  30. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023).
  31. Conversational recommendation via hierarchical information modeling. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2201–2205.
  32. Attention is all you need. Advances in neural information processing systems 30 (2017).
  33. Ellen M Voorhees and L Buckland. 2003. Overview of the TREC 2003 Question Answering Track.. In TREC, Vol. 2003. 54–68.
  34. Lei Wang and Ee-Peng Lim. 2023. Zero-Shot Next-Item Recommendation using Large Pretrained Language Models. arXiv preprint arXiv:2304.03153 (2023).
  35. Rethinking the Evaluation for Conversational Recommendation in the Era of Large Language Models. arXiv preprint arXiv:2305.13112 (2023).
  36. A Survey on Large Language Models for Recommendation. arXiv preprint arXiv:2305.19860 (2023).
  37. Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-Based Chatbots. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 496–505.
  38. Glm-130b: An open bilingual pre-trained model. ICLR (2022).
  39. Is chatgpt fair for recommendation? evaluating fairness in large language model recommendation. arXiv preprint arXiv:2305.07609 (2023).
  40. Recommendation as instruction following: A large language model empowered recommendation approach. arXiv preprint arXiv:2305.07001 (2023).
  41. Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022).
  42. CRFR: Improving conversational recommender systems via flexible fragments reasoning on knowledge graphs. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. 4324–4334.
  43. CRSLab: An Open-Source Toolkit for Building Conversational Recommender System. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics: System Demonstrations. 185–193.
  44. Improving conversational recommender systems via knowledge graph based semantic fusion. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining. 1006–1014.
  45. Leveraging historical interaction data for improving conversational recommender system. In Proceedings of the 29th ACM international conference on information & knowledge management. 2349–2352.
  46. Towards Topic-Guided Conversational Recommender System. In Proceedings of the 28th International Conference on Computational Linguistics. 4128–4139.
  47. C22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT-CRS: Coarse-to-Fine Contrastive Learning for Conversational Recommender System. In Proceedings of WSDM. 1488–1496.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yue Feng (55 papers)
  2. Shuchang Liu (39 papers)
  3. Zhenghai Xue (13 papers)
  4. Qingpeng Cai (43 papers)
  5. Lantao Hu (20 papers)
  6. Peng Jiang (272 papers)
  7. Kun Gai (125 papers)
  8. Fei Sun (151 papers)
Citations (23)