A Large Language Model Enhanced Conversational Recommender System (2308.06212v1)
Abstract: Conversational recommender systems (CRSs) aim to recommend high-quality items to users through a dialogue interface. It usually contains multiple sub-tasks, such as user preference elicitation, recommendation, explanation, and item information search. To develop effective CRSs, there are some challenges: 1) how to properly manage sub-tasks; 2) how to effectively solve different sub-tasks; and 3) how to correctly generate responses that interact with users. Recently, LLMs have exhibited an unprecedented ability to reason and generate, presenting a new opportunity to develop more powerful CRSs. In this work, we propose a new LLM-based CRS, referred to as LLMCRS, to address the above challenges. For sub-task management, we leverage the reasoning ability of LLM to effectively manage sub-task. For sub-task solving, we collaborate LLM with expert models of different sub-tasks to achieve the enhanced performance. For response generation, we utilize the generation ability of LLM as a language interface to better interact with users. Specifically, LLMCRS divides the workflow into four stages: sub-task detection, model matching, sub-task execution, and response generation. LLMCRS also designs schema-based instruction, demonstration-based instruction, dynamic sub-task and model matching, and summary-based generation to instruct LLM to generate desired results in the workflow. Finally, to adapt LLM to conversational recommendations, we also propose to fine-tune LLM with reinforcement learning from CRSs performance feedback, referred to as RLPF. Experimental results on benchmark datasets show that LLMCRS with RLPF outperforms the existing methods.
- Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
- A survey on dialogue systems: Recent advances and new frontiers. Acm Sigkdd Explorations Newsletter 19, 2 (2017), 25–35.
- Towards Knowledge-Based Recommender Dialog System. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. 1803–1813.
- Palm: Scaling language modeling with pathways. arXiv preprint arXiv:2204.02311 (2022).
- Scaling instruction-finetuned language models. arXiv preprint arXiv:2210.11416 (2022).
- Uncovering ChatGPT’s Capabilities in Recommender Systems. arXiv preprint arXiv:2305.02182 (2023).
- Advances and challenges in conversational recommender systems: A survey. AI Open 2 (2021), 100–126.
- Neural approaches to conversational AI. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 1371–1374.
- Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv preprint arXiv:2303.14524 (2023).
- Kalervo Järvelin and Jaana Kekäläinen. 2017. IR evaluation methods for retrieving highly relevant documents. In ACM SIGIR Forum, Vol. 51. 243–250.
- Recommendation as a Communication Game: Self-Supervised Bot-Play for Goal-oriented Dialogue. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. 1951–1961.
- Jacob Devlin Ming-Wei Chang Kenton and Lee Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of NAACL-HLT. 4171–4186.
- Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
- Estimation-action-reflection: Towards deep interaction between conversational and recommender systems. In Proceedings of the 13th International Conference on Web Search and Data Mining. 304–312.
- A Diversity-Promoting Objective Function for Neural Conversation Models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 110–119.
- Towards deep conversational recommendations. Advances in neural information processing systems 31 (2018).
- Yuxi Li. 2017. Deep reinforcement learning: An overview. arXiv preprint arXiv:1701.07274 (2017).
- Pre-train, prompt and recommendation: A comprehensive survey of language modelling paradigm adaptations in recommender systems. arXiv preprint arXiv:2302.03735 (2023).
- The flan collection: Designing data and methods for effective instruction tuning. arXiv preprint arXiv:2301.13688 (2023).
- ParlAI: A Dialog Research Software Platform. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. 79–84.
- Key-Value Memory Networks for Directly Reading Documents. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 1400–1409.
- Top-k learning to rank: labeling, ranking and evaluation. In Proceedings of SIGIR. 751–760.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems 35 (2022), 27730–27744.
- Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics. 311–318.
- Language models are unsupervised multitask learners. (2018).
- Toolformer: Language models can teach themselves to use tools. arXiv preprint arXiv:2302.04761 (2023).
- Multi-grained Hypergraph Interest Modeling for Conversational Recommendation. arXiv preprint arXiv:2305.04798 (2023).
- Hugginggpt: Solving ai tasks with chatgpt and its friends in huggingface. arXiv preprint arXiv:2303.17580 (2023).
- Richard S Sutton and Andrew G Barto. 2018. Reinforcement learning: An introduction. MIT press.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023).
- Conversational recommendation via hierarchical information modeling. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval. 2201–2205.
- Attention is all you need. Advances in neural information processing systems 30 (2017).
- Ellen M Voorhees and L Buckland. 2003. Overview of the TREC 2003 Question Answering Track.. In TREC, Vol. 2003. 54–68.
- Lei Wang and Ee-Peng Lim. 2023. Zero-Shot Next-Item Recommendation using Large Pretrained Language Models. arXiv preprint arXiv:2304.03153 (2023).
- Rethinking the Evaluation for Conversational Recommendation in the Era of Large Language Models. arXiv preprint arXiv:2305.13112 (2023).
- A Survey on Large Language Models for Recommendation. arXiv preprint arXiv:2305.19860 (2023).
- Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-Based Chatbots. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 496–505.
- Glm-130b: An open bilingual pre-trained model. ICLR (2022).
- Is chatgpt fair for recommendation? evaluating fairness in large language model recommendation. arXiv preprint arXiv:2305.07609 (2023).
- Recommendation as instruction following: A large language model empowered recommendation approach. arXiv preprint arXiv:2305.07001 (2023).
- Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068 (2022).
- CRFR: Improving conversational recommender systems via flexible fragments reasoning on knowledge graphs. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing. 4324–4334.
- CRSLab: An Open-Source Toolkit for Building Conversational Recommender System. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics: System Demonstrations. 185–193.
- Improving conversational recommender systems via knowledge graph based semantic fusion. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining. 1006–1014.
- Leveraging historical interaction data for improving conversational recommender system. In Proceedings of the 29th ACM international conference on information & knowledge management. 2349–2352.
- Towards Topic-Guided Conversational Recommender System. In Proceedings of the 28th International Conference on Computational Linguistics. 4128–4139.
- C22{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT-CRS: Coarse-to-Fine Contrastive Learning for Conversational Recommender System. In Proceedings of WSDM. 1488–1496.
- Yue Feng (55 papers)
- Shuchang Liu (39 papers)
- Zhenghai Xue (13 papers)
- Qingpeng Cai (43 papers)
- Lantao Hu (20 papers)
- Peng Jiang (272 papers)
- Kun Gai (125 papers)
- Fei Sun (151 papers)