Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bridging Items and Language: A Transition Paradigm for Large Language Model-Based Recommendation (2310.06491v2)

Published 10 Oct 2023 in cs.IR

Abstract: Harnessing LLMs for recommendation is rapidly emerging, which relies on two fundamental steps to bridge the recommendation item space and the language space: 1) item indexing utilizes identifiers to represent items in the language space, and 2) generation grounding associates LLMs' generated token sequences to in-corpus items. However, previous methods exhibit inherent limitations in the two steps. Existing ID-based identifiers (e.g., numeric IDs) and description-based identifiers (e.g., titles) either lose semantics or lack adequate distinctiveness. Moreover, prior generation grounding methods might generate invalid identifiers, thus misaligning with in-corpus items. To address these issues, we propose a novel Transition paradigm for LLM-based Recommender (named TransRec) to bridge items and language. Specifically, TransRec presents multi-facet identifiers, which simultaneously incorporate ID, title, and attribute for item indexing to pursue both distinctiveness and semantics. Additionally, we introduce a specialized data structure for TransRec to ensure generating valid identifiers only and utilize substring indexing to encourage LLMs to generate from any position of identifiers. Lastly, TransRec presents an aggregated grounding module to leverage generated multi-facet identifiers to rank in-corpus items efficiently. We instantiate TransRec on two backbone models, BART-large and LLaMA-7B. Extensive results on three real-world datasets under diverse settings validate the superiority of TransRec.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. A bi-step grounding paradigm for large language models in recommendation systems. arXiv:2308.08434.
  2. Tallrec: An effective and efficient tuning framework to align large language model with recommendation. In RecSys. ACM.
  3. Towards language models that can see: Computer vision through the lens of natural language. arXiv:2306.16410.
  4. Autoregressive search engines: Generating substrings as document identifiers. NeurIPS 35 (2022), 31668–31683.
  5. Zheng Chen. 2023. PALR: Personalization Aware LLMs for Recommendation. arXiv preprint arXiv:2305.07622 (2023).
  6. M6-rec: Generative pretrained language models are open-ended recommender systems. arXiv:2205.08084.
  7. Uncovering ChatGPT’s Capabilities in Recommender Systems. arXiv:2305.02182.
  8. Autoregressive entity retrieval. arXiv:2010.00904.
  9. Toward personalized answer generation in e-commerce via multi-perspective preference modeling. TOIS 40, 4 (2022), 1–28.
  10. Paolo Ferragina and Giovanni Manzini. 2000. Opportunistic data structures with applications. In Proceedings 41st annual symposium on foundations of computer science. IEEE, 390–398.
  11. Chat-rec: Towards interactive and explainable llms-augmented recommender system. arXiv:2303.14524.
  12. Recommendation as language processing (rlp): A unified pretrain, personalized prompt & predict paradigm (p5). In RecSys. ACM, 299–315.
  13. Retrieval augmented language model pre-training. In ICML. PMLR, 3929–3938.
  14. Lightgcn: Simplifying and powering graph convolution network for recommendation. In SIGIR. 639–648.
  15. Session-based recommendations with recurrent neural networks. In ICLR.
  16. Bros: A pre-trained language model focusing on text and layout for better key information extraction from documents. In AAAI, Vol. 36. AAAI press, 10767–10775.
  17. Large language models are zero-shot rankers for recommender systems. arXiv:2305.08845.
  18. Lora: Low-rank adaptation of large language models. arXiv:2106.09685.
  19. How to Index Item IDs for Recommendation Foundation Models. arXiv:2305.06569.
  20. Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In ICDM. IEEE, 197–206.
  21. Diederik P Kingma and Jimmy Ba. 2014. Adam: A Method for Stochastic Optimization. In arXiv:1412.6980.
  22. Large language models are zero-shot reasoners. NeurIPS 35 (2022), 22199–22213.
  23. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv:1910.13461.
  24. Multiview Identifiers Enhanced Generative Retrieval. In ACL. ACM.
  25. Code as policies: Language model programs for embodied control. In ICRA. IEEE, 9493–9500.
  26. Is chatgpt a good recommender? a preliminary study. arXiv:2304.10149.
  27. Ilya Loshchilov and Frank Hutter. 2017. Decoupled weight decay regularization. arXiv:1711.05101.
  28. Training language models to follow instructions with human feedback. NeurIPS 35 (2022), 27730–27744.
  29. BPR: Bayesian personalized ranking from implicit feedback. In UAI. AUAI Press, 452–461.
  30. Llama: Open and efficient foundation language models. arXiv:2302.13971.
  31. Lei Wang and Ee-Peng Lim. 2023. Zero-Shot Next-Item Recommendation using Large Pretrained Language Models. arXiv:2304.03153.
  32. Mm-rec: Visiolinguistic model empowered multimodal news recommendation. In SIGIR. ACM, 2560–2564.
  33. Towards Open-World Recommendation with Knowledge Augmentation from Large Language Models. arXiv:2306.10933.
  34. Adversarial and contrastive variational autoencoder for sequential recommendation. In WWW. ACM, 449–459.
  35. Recommendation as instruction following: A large language model empowered recommendation approach. arXiv:2305.07001.
  36. UNBERT: User-News Matching BERT for News Recommendation.. In IJCAI. 3356–3362.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Xinyu Lin (24 papers)
  2. Wenjie Wang (150 papers)
  3. Yongqi Li (40 papers)
  4. Fuli Feng (143 papers)
  5. See-Kiong Ng (103 papers)
  6. Tat-Seng Chua (359 papers)
Citations (28)