Play to Your Strengths: Collaborative Intelligence of Conventional Recommender Models and Large Language Models (2403.16378v1)
Abstract: The rise of LLMs has opened new opportunities in Recommender Systems (RSs) by enhancing user behavior modeling and content understanding. However, current approaches that integrate LLMs into RSs solely utilize either LLM or conventional recommender model (CRM) to generate final recommendations, without considering which data segments LLM or CRM excel in. To fill in this gap, we conduct experiments on MovieLens-1M and Amazon-Books datasets, and compare the performance of a representative CRM (DCNv2) and an LLM (LLaMA2-7B) on various groups of data samples. Our findings reveal that LLMs excel in data segments where CRMs exhibit lower confidence and precision, while samples where CRM excels are relatively challenging for LLM, requiring substantial training data and a long training time for comparable performance. This suggests potential synergies in the combination between LLM and CRM. Motivated by these insights, we propose Collaborative Recommendation with conventional Recommender and LLM (dubbed \textit{CoReLLa}). In this framework, we first jointly train LLM and CRM and address the issue of decision boundary shifts through alignment loss. Then, the resource-efficient CRM, with a shorter inference time, handles simple and moderate samples, while LLM processes the small subset of challenging samples for CRM. Our experimental results demonstrate that CoReLLa outperforms state-of-the-art CRM and LLM methods significantly, underscoring its effectiveness in recommendation tasks.
- Tallrec: An effective and efficient tuning framework to align large language model with recommendation. arXiv preprint arXiv:2305.00447, 2023.
- Being confident about the quality of the predictions in recommender systems. In ECIR, pages 411–422. Springer, 2013.
- M6-rec: Generative pretrained language models are open-ended recommender systems. arXiv preprint arXiv:2205.08084, 2022.
- Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 320–335, 2022.
- Leveraging large language models in conversational recommender systems. arXiv preprint arXiv:2305.07961, 2023.
- Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv preprint arXiv:2303.14524, 2023.
- Recommendation as language processing (rlp): A unified pretrain, personalized prompt and predict paradigm (p5). In RecSys, page 299–315, 2022.
- An unified search and recommendation foundation model for cold-start scenario. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, CIKM ’23, page 4595–4601, 2023.
- Habituation: a dual-process theory. Psychological review, 77(5):419, 1970.
- Large language models are zero-shot rankers for recommender systems. arXiv preprint arXiv:2305.08845, 2023.
- Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685, 2021.
- Fibinet: Combining feature importance and bilinear feature interaction for click-through rate prediction. In RecSys, page 169–177, 2019.
- Improved recommender systems by denoising ratings in highly sparse datasets through individual rating confidence. Information Sciences, 601:242–254, 2022.
- Ctrl: Connect tabular and language model for ctr prediction. arXiv preprint arXiv:2306.02841, 2023.
- Fi-gnn: Modeling feature interactions via graph neural networks for ctr prediction. In CIKM, page 539–548, 2019.
- Xdeepfm: Combining explicit and implicit feature interactions for recommender systems. In KDD, page 1754–1763, 2018.
- Clickprompt: Ctr models are strong prompt generators for adapting language models to ctr prediction. arXiv preprint arXiv:2310.09234, 2023a.
- How can recommender systems benefit from large language models: A survey. arXiv preprint arXiv:2306.05817, 2023b.
- Map: A model-agnostic pretraining framework for click-through rate prediction. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pages 1384–1395, 2023c.
- Rella: Retrieval-enhanced large language models for lifelong sequential behavior comprehension in recommendation. arXiv preprint arXiv:2308.11131, 2023d.
- Is chatgpt a good recommender? a preliminary study. arXiv preprint arXiv:2304.10149, 2023.
- Recranker: Instruction tuning large language model as ranker for top-k recommendation. arXiv preprint arXiv:2312.16018, 2023.
- Llm-rec: Personalized recommendation via prompting large language models. arXiv preprint arXiv:2307.15780, 2023.
- Justifying recommendations using distantly-labeled reviews and fine-grained aspects. In Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pages 188–197, 2019.
- Laurence Anthony F. Park and Simeon Simoff. Using entropy as a measure of acceptance for multi-label classification. In XIV, 2015.
- User behavior retrieval for click-through rate prediction. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2347–2356, 2020.
- Large language models are competitive near cold-start recommenders for language-and item-based preferences. In Proceedings of the 17th ACM conference on recommender systems, pages 890–896, 2023.
- Autoint: Automatic feature interaction learning via self-attentive neural networks. In CIKM, page 1161–1170, 2019.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023a.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023b.
- Flip: Towards fine-grained alignment between id-based models and pretrained language models for ctr prediction. arXiv e-prints, pages arXiv–2310, 2023a.
- Dcn v2: Improved deep & cross network and practical lessons for web-scale learning to rank systems. In WWW, page 1785–1797, 2021.
- Augmentation with projection: Towards an effective and efficient data augmentation paradigm for distillation. In ICML, 2023b.
- Towards open-world recommendation with knowledge augmentation from large language models. arXiv preprint arXiv:2306.10933, 2023.
- Language models as recommender systems: Evaluations and limitations. In I (Still) Can’t Believe It’s Not Better! NeurIPS 2021 Workshop, 2021.
- Deep interest network for click-through rate prediction. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, page 1059–1068, 2018.
- Yunjia Xi (21 papers)
- Weiwen Liu (59 papers)
- Jianghao Lin (47 papers)
- Chuhan Wu (86 papers)
- Bo Chen (309 papers)
- Ruiming Tang (171 papers)
- Weinan Zhang (322 papers)
- Yong Yu (219 papers)