Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RecGPT: Generative Pre-training for Text-based Recommendation (2405.12715v1)

Published 21 May 2024 in cs.IR and cs.CL

Abstract: We present the first domain-adapted and fully-trained LLM, RecGPT-7B, and its instruction-following variant, RecGPT-7B-Instruct, for text-based recommendation. Experimental results on rating prediction and sequential recommendation tasks show that our model, RecGPT-7B-Instruct, outperforms previous strong baselines. We are releasing our RecGPT models as well as their pre-training and fine-tuning datasets to facilitate future research and downstream applications in text-based recommendation. Public "huggingface" links to our RecGPT models and datasets are available at: https://github.com/VinAIResearch/RecGPT

RecGPT: A Text-Centric Approach to Recommendation Systems

Overview

RecGPT-7B is a newly-developed LLM designed specifically for text-based recommendation tasks. It focuses on two key areas: rating prediction and sequential recommendation. The research team behind RecGPT-7B has demonstrated that it outperforms existing models like P5 and even some versions of ChatGPT in various evaluations. They have made the model, along with the pre-training and fine-tuning datasets, publicly available, which could be useful for both researchers and developers working on recommendation systems.

Introduction and Background

Recommendation systems are essential in curating user experiences across platforms such as e-commerce, news, and streaming services. Traditional methods often rely on interaction matrices and deep learning architectures to predict user preferences. However, these approaches face challenges related to data sparsity and adaptability over time.

In recent years, leveraging LLMs in recommendation tasks has emerged as a promising alternative. Models like P5 have shown that representing users and items in text can better aggregate recommendation tasks. Building upon this, RecGPT-7B aims to improve on such models by incorporating instructional fine-tuning and domain-specific data.

The RecGPT Model

Pre-training and Fine-tuning

Data Collection:

The team collected a massive dataset from various sources, including Amazon, Yelp, Goodreads, and Netflix. They ensured data quality by discarding incomplete entries and irrelevant information. This process resulted in a dataset with over 10 million users and 10 million items, yielding around 258 million interactions.

Pre-training:

RecGPT-7B was initially trained on a corpus of 20.5 billion tokens. The training utilized advanced techniques like flash attention and ALiBi for better handling long context sequences. The model training spanned 18 days using 8 A100 GPUs, emphasizing a rigorous, resource-intensive process.

Fine-tuning:

Fine-tuning was performed on 100,000+ prompts and responses related to rating prediction and sequential recommendation. This step adapted the model to instruction-following tasks, enhancing its ability to generate relevant predictions based on textual inputs.

Key Findings

Rating Prediction

RecGPT-7B-Instruct showcased impressive performance in rating prediction across several datasets like Amazon Beauty and Yelp. The model achieved lower Root Mean Square Error (RMSE) and Mean Absolute Error (MAE) compared to competitors like P5, setting new benchmarks.

Sequential Recommendation

For sequential recommendation tasks, RecGPT-7B-Instruct also delivered competitive results. It managed to outperform P5 in several instances, such as the "Sports and Outdoors" and "Toys and Games" datasets.

Ablation Analysis

To understand the advantages of the domain-specific pre-training, the team compared RecGPT-7B-Instruct with a variant trained solely on the base pre-trained MPT-7B. The former showed significantly better results, highlighting the importance of specialized pre-training for recommendation tasks.

Implications and Future Directions

Practical Implications:

The superior performance of RecGPT-7B-Instruct suggests that LLMs customized for text-based recommendations can more accurately predict user preferences. This can lead to better user experiences in platforms that rely on personalized recommendations, such as e-commerce and entertainment services.

Theoretical Implications:

The paper underscores the potential for domain-specific adaptations of LLMs to solve specialized tasks. This approach could pave the way for developing more efficient and effective models across various domains beyond recommendations.

Future Directions:

Given the promising results, future work can expand to other types of recommendation tasks. Additionally, mitigation strategies for model hallucinations and adaptation to dynamic item sets could be explored. The public release of RecGPT-7B and its datasets also opens avenues for collaborative research and practical deployments.

Conclusion

RecGPT-7B represents a significant step forward in the field of text-based recommendation systems. Its robust performance across multiple datasets validates the potential of LLMs customized for specific domains. With the public release of the model and its datasets, the doors are open for further innovation and application in the field of recommendation systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Hoang Ngo (3 papers)
  2. Dat Quoc Nguyen (55 papers)
Citations (3)