Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Zero-Shot Next-Item Recommendation using Large Pretrained Language Models (2304.03153v1)

Published 6 Apr 2023 in cs.IR and cs.CL
Zero-Shot Next-Item Recommendation using Large Pretrained Language Models

Abstract: LLMs have achieved impressive zero-shot performance in various NLP tasks, demonstrating their capabilities for inference without training examples. Despite their success, no research has yet explored the potential of LLMs to perform next-item recommendations in the zero-shot setting. We have identified two major challenges that must be addressed to enable LLMs to act effectively as recommenders. First, the recommendation space can be extremely large for LLMs, and LLMs do not know about the target user's past interacted items and preferences. To address this gap, we propose a prompting strategy called Zero-Shot Next-Item Recommendation (NIR) prompting that directs LLMs to make next-item recommendations. Specifically, the NIR-based strategy involves using an external module to generate candidate items based on user-filtering or item-filtering. Our strategy incorporates a 3-step prompting that guides GPT-3 to carry subtasks that capture the user's preferences, select representative previously watched movies, and recommend a ranked list of 10 movies. We evaluate the proposed approach using GPT-3 on MovieLens 100K dataset and show that it achieves strong zero-shot performance, even outperforming some strong sequential recommendation models trained on the entire training dataset. These promising results highlight the ample research opportunities to use LLMs as recommenders. The code can be found at https://github.com/AGI-Edgerunners/LLM-Next-Item-Rec.

Zero-Shot Next-Item Recommendation Using Large Pretrained LLMs

The paper "Zero-Shot Next-Item Recommendation using Large Pretrained LLMs" presents a novel approach that leverages the capabilities of LLMs to address a challenging problem in recommender systems: making next-item recommendations without any example demonstration or additional training. This research systematically explores the applicability of LLMs to perform zero-shot next-item recommendations, specifically leveraging GPT-3, a state-of-the-art LLM.

Core Contributions

The paper identifies two primary obstacles when applying LLMs to recommendation tasks: the potential enormity of the recommendation space and the lack of user-specific historical interaction data. To address these challenges, the authors propose a "Zero-Shot Next-Item Recommendation" (NIR) prompting strategy, utilizing a multi-step approach fundamentally different from simplistic single-prompt strategies previously explored.

  1. Candidate Set Construction: This process dramatically narrows the recommendation space by selecting a subset of potentially relevant items for each user using user filtering and item filtering. By representing users and items as vectors and calculating cosine similarities, the candidate set is determined based on popularity and similarity metrics, ensuring this subset is manageable in size while relevant to the user.
  2. Three-Step Prompting Strategy: This strategy encompasses a sequence of prompts designed to guide GPT-3 through capturing user preferences, identifying representative user-interacted items that best reflect these preferences, and finally recommending new items from the candidate space. Each step leverages the intermediate results to fine-tune the subsequent prompt, effectively guiding the model's reasoning process.
  3. Answer Extraction: To handle GPT-3's output, a format-based method is utilized, ensuring that the model's textual recommendations can be efficiently parsed and integrated into the recommendation system.

Empirical Evaluation

The proposed method's effectiveness was tested on the MovieLens 100K dataset, and it demonstrated significant promise as a zero-shot recommendation approach. Notably, the NIR methods saw performance metrics such as HR@10 and NDCG@10 showing competitive results not only against simple zero-shot prompting baselines but even surpassing certain models that were fully trained on the dataset. This highlights NIR's potential in scenarios where training data is scarce or non-existent.

Furthermore, the careful design of a principled strategy to reduce the candidate item set, along with the multi-step prompting, was shown to improve performance over simple or random selection methods significantly. Additionally, the impact of tuning the candidate set size was explored, evidencing the trade-off between candidate set comprehensiveness and recommendation precision.

Implications and Future Directions

By demonstrating that LLMs can perform effectively in zero-shot recommendation tasks, this research potentially shifts how recommender systems can be designed, especially in context-rich domains where user-specific training data might not be feasible to collect.

Future work could extend this methodology across different domains and explore how richer interactions (beyond simple watch history) can be incorporated into LLM prompts. Additionally, investigating few-shot learning scenarios could offer a productive middle ground between zero-shot settings and full model training, providing pathways for more nuanced and customizable user experiences.

In conclusion, this research provides a compelling look into the adaptability of LLMs beyond traditional NLP tasks, showcasing their potential in operational domains like recommender systems without necessitating time-consuming pretraining or fine-tuning processes. The implications for practical applications are manifold, laying groundwork for more flexible, responsive, and less data-intensive systems across industries reliant on recommendation algorithms.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Lei Wang (975 papers)
  2. Ee-Peng Lim (57 papers)
Citations (47)