Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ChatGPT for Conversational Recommendation: Refining Recommendations by Reprompting with Feedback (2401.03605v1)

Published 7 Jan 2024 in cs.IR, cs.AI, cs.CL, and cs.LG

Abstract: Recommendation algorithms have been pivotal in handling the overwhelming volume of online content. However, these algorithms seldom consider direct user input, resulting in superficial interaction between them. Efforts have been made to include the user directly in the recommendation process through conversation, but these systems too have had limited interactivity. Recently, LLMs like ChatGPT have gained popularity due to their ease of use and their ability to adapt dynamically to various tasks while responding to feedback. In this paper, we investigate the effectiveness of ChatGPT as a top-n conversational recommendation system. We build a rigorous pipeline around ChatGPT to simulate how a user might realistically probe the model for recommendations: by first instructing and then reprompting with feedback to refine a set of recommendations. We further explore the effect of popularity bias in ChatGPT's recommendations, and compare its performance to baseline models. We find that reprompting ChatGPT with feedback is an effective strategy to improve recommendation relevancy, and that popularity bias can be mitigated through prompt engineering.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877–1901.
  2. Second workshop on information heterogeneity and fusion in recommender systems (HetRec2011). In Proceedings of the fifth ACM conference on Recommender systems. 387–388.
  3. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
  4. Complexity-based prompting for multi-step reasoning. arXiv preprint arXiv:2210.00720 (2022).
  5. Chat-REC: Towards Interactive and Explainable LLMs-Augmented Recommender System. arXiv preprint arXiv:2303.14524 (03 2023).
  6. Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv preprint arXiv:2303.14524 (2023).
  7. F Maxwell Harper and Joseph A Konstan. 2015. The movielens datasets: History and context. Acm transactions on interactive intelligent systems (tiis) 5, 4 (2015), 1–19.
  8. An algorithmic framework for performing collaborative filtering. In Proceedings of the 22nd annual international ACM SIGIR conference on Research and development in information retrieval. 230–237.
  9. Introducing serendipity in a content-based recommender system. In 2008 eighth international conference on hybrid intelligent systems. IEEE, 168–173.
  10. IMDb. 2023. IMDb Top 250. https://www.imdb.com/chart/top/ Accessed: 2023-07.
  11. Intra-list similarity and human diversity perceptions of recommendations: the details matter. User Modeling and User-Adapted Interaction (2022), 1–34.
  12. Marius Kaminskas and Derek Bridge. 2016. Diversity, serendipity, novelty, and coverage: a survey and empirical analysis of beyond-accuracy objectives in recommender systems. ACM Transactions on Interactive Intelligent Systems (TiiS) 7, 1 (2016), 1–42.
  13. Do LLMs Understand User Preferences? Evaluating LLMs On User Rating Prediction. arXiv preprint arXiv:2305.06474 (2023).
  14. A Survey of Algorithmic Recourse: Contrastive Explanations and Consequential Recommendations. ACM Comput. Surv. 55, 5, Article 95 (dec 2022), 29 pages. https://doi.org/10.1145/3527848
  15. Bing Liu et al. 2011. Web data mining: exploring hyperlinks, contents, and usage data. Vol. 1. Springer.
  16. Is chatgpt a good recommender? a preliminary study. arXiv preprint arXiv:2304.10149 (2023).
  17. Training language models to follow instructions with human feedback. arXiv:2203.02155 [cs.CL]
  18. Language models are unsupervised multitask learners. OpenAI blog 1, 8 (2019), 9.
  19. Large Language Models are Competitive Near Cold-start Recommenders for Language- and Item-based Preferences. arXiv:2307.14225 [cs.IR]
  20. Item-based collaborative filtering recommendation algorithms. In Proceedings of the 10th international conference on World Wide Web. 285–295.
  21. BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer. arXiv:1904.06690 [cs.IR]
  22. A personalized system for conversational recommendations. Journal of Artificial Intelligence Research 21 (2004), 393–428.
  23. Actionable Recourse in Linear Classification. In Proceedings of the Conference on Fairness, Accountability, and Transparency. ACM. https://doi.org/10.1145/3287560.3287566
  24. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022), 24824–24837.
  25. Xi Ye and Greg Durrett. 2022. The unreliability of explanations in few-shot in-context learning. arXiv preprint arXiv:2205.03401 (2022).
  26. Li Yujian and Liu Bo. 2007. A normalized Levenshtein distance metric. IEEE transactions on pattern analysis and machine intelligence 29, 6 (2007), 1091–1095.
  27. Language models as recommender systems: Evaluations and limitations. (2021).
  28. Improving recommendation lists through topic diversification. In Proceedings of the 14th international conference on World Wide Web. 22–32.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Kyle Dylan Spurlock (1 paper)
  2. Cagla Acun (1 paper)
  3. Esin Saka (1 paper)
  4. Olfa Nasraoui (10 papers)
Citations (9)
X Twitter Logo Streamline Icon: https://streamlinehq.com