Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RecJPQ: Training Large-Catalogue Sequential Recommenders (2312.06165v2)

Published 11 Dec 2023 in cs.IR

Abstract: Sequential Recommendation is a popular recommendation task that uses the order of user-item interaction to model evolving users' interests and sequential patterns in their behaviour. Current state-of-the-art Transformer-based models for sequential recommendation, such as BERT4Rec and SASRec, generate sequence embeddings and compute scores for catalogue items, but the increasing catalogue size makes training these models costly. The Joint Product Quantisation (JPQ) method, originally proposed for passage retrieval, markedly reduces the size of the retrieval index with minimal effect on model effectiveness, by replacing passage embeddings with a limited number of shared sub-embeddings. This paper introduces RecJPQ, a novel adaptation of JPQ for sequential recommendations, which takes the place of item embeddings tensor and replaces item embeddings with a concatenation of a limited number of shared sub-embeddings and, therefore, limits the number of learnable model parameters. The main idea of RecJPQ is to split items into sub-item entities before training the main recommendation model, which is inspired by splitting words into tokens and training tokenisers in LLMs. We apply RecJPQ to SASRec, BERT4Rec, and GRU4rec models on three large-scale sequential datasets. Our results showed that RecJPQ could notably reduce the model size (e.g., 48% reduction for the Gowalla dataset with no effectiveness degradation). RecJPQ can also improve model performance through a regularisation effect (e.g. +0.96% NDCG@10 improvement on the Booking.com dataset). Overall, RecJPQ allows the training of state-of-the-art transformer recommenders in industrial applications, where datasets with millions of items are common.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (66)
  1. 2023. Dimensionality Reduction - RDD-based API. https://spark.apache.org/docs/latest/mllib-dimensionality-reduction. [Online; accessed 14-December-2023].
  2. 2023. How Many Videos Are on YouTube in 2024? https://earthweb.com/how-many-videos-are-on-youtube/. [Online; accessed 14-December-2023].
  3. Jan Van Balen and Mark Levy. 2019. PQ-VAE: Efficient Recommendation Using Quantized Embeddings. In Proc. RecSys. 46–50.
  4. Language Models Are Few-Shot Learners. In Proc. NeurIPS, Vol. 33. 1877–1901.
  5. Christopher Burges. 2010. From RankNet to LambdaRank to LambdaMART: An overview. Learning 11 (2010).
  6. Rocío Cañamares and Pablo Castells. 2020. On Target Item Sampling in Offline Recommender System Evaluation. In Proc. RecSys. 259–268.
  7. Friendship and mobility: user movement in location-based social networks. In Proc. KDD. 1082–1090.
  8. Kyunghyun Cho and Bart van Merrienboer. 2014. On the Properties of Neural Machine Translation: Encoder-Decoder Approaches. (2014). arXiv:1409.1259 [cs, stat]
  9. A Case Study on Sampling Strategies for Evaluating Neural Sequential Item Recommendation Models. In Proc. RecSys. 505–514.
  10. Kalyanmoy Deb and Kalyanmoy Deb. 2014. Multi-objective Optimization. In Search Methodologies: Introductory Tutorials in Optimization and Decision Support Techniques. 403–449.
  11. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proc. NAACL-HLT. 4171–4186.
  12. Optimized Product Quantization. IEEE Transactions on Pattern Analysis and Machine Intelligence 36, 4 (2014), 744–755.
  13. Dmitri Goldenberg and Pavel Levin. 2021. Booking.com Multi-Destination Trips Dataset. In Proc. SIGIR. 2457–2462.
  14. Deep Learning. MIT Press.
  15. Robert M. Gray. 1984. Vector Quantization. IEEE Assp 1, 2 (1984).
  16. Finding Structure with Randomness: Probabilistic Algorithms for Constructing Approximate Matrix Decompositions. 53, 2 (2011), 217–288.
  17. F. Maxwell Harper and Joseph A. Konstan. 2015. The MovieLens Datasets: History and Context. 5, 4 (2015), 19:1–19:19.
  18. Neural Collaborative Filtering. In Proc. WWW. 173–182.
  19. Balázs Hidasi and Alexandros Karatzoglou. 2018. Recurrent Neural Networks with Top-k Gains for Session-based Recommendations. In Proc. CIKM. 843–852.
  20. Session-Based Recommendations with Recurrent Neural Networks. In Proc. ICLR.
  21. Learning Vector-Quantized Item Representation for Transferable Sequential Recommenders. In Proc. WWW. 1162–1171.
  22. Learning Deep Structured Semantic Models for Web Search Using Clickthrough Data. In Proc. CIKM. 2333–2338.
  23. A Memory Efficient Baseline for Open Domain Question Answering. arXiv:2012.15156 [cs]
  24. Product Quantization for Nearest Neighbor Search. IEEE Transactions on Pattern Analysis and Machine Intelligence 33, 1 (2011), 117–128.
  25. Billion-Scale Similarity Search with GPUs. IEEE Transactions on Big Data 7, 3 (2021), 535–547.
  26. Learning Multi-granular Quantized Embeddings for Large-Vocab Categorical Features in Recommender Systems. In Proc. WWW. 562–566.
  27. Wang-Cheng Kang and Julian McAuley. 2018. Self-Attentive Sequential Recommendation. In Proc. ICDM. 197–206.
  28. Dense Passage Retrieval for Open-Domain Question Answering. In Proc. EMNLP. arXiv:2004.04906 [cs]
  29. Omar Khattab and Matei Zaharia. 2020. ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT. In Proc. SIGIR. 39–48.
  30. Matrix Factorization Techniques for Recommender Systems. Computer 42, 8 (2009), 30–37.
  31. Walid Krichene and Steffen Rendle. 2022. On sampled metrics for item recommendation. Commun. ACM 65, 7 (2022), 75–83.
  32. Approximate Nearest Neighbor Search on High Dimensional Data — Experiments, Analyses, and Improvement. IEEE Transactions on Knowledge and Data Engineering 32, 8 (2020), 1475–1488.
  33. LightRec: A Memory and Search-Efficient Recommender System. In Proc. WWW. 695–705.
  34. Alfredo Láinez Luke de Oliveira. 2018. Bayesian Personalized Ranking for Spark. https://github.com/alfredolainez/bpr-spark. [Online; accessed 14-December-2023].
  35. J. MacQueen. 1967. Some Methods for Classification and Analysis of Multivariate Observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Statistics. Vol. 5.1. 281–298.
  36. Yu A. Malkov and D. A. Yashunin. 2020. Efficient and Robust Approximate Nearest Neighbor Search Using Hierarchical Navigable Small World Graphs. IEEE Transactions on Pattern Analysis and Machine Intelligence 42, 4 (2020), 824–836.
  37. Mixed Precision Training. In Proc. ICLR.
  38. Deep Learning Recommendation Model for Personalization and Recommendation Systems. arXiv:1906.00091 [cs]
  39. Yoon-Joo Park and Alexander Tuzhilin. 2008. The long tail of recommender systems and how to leverage it. In Proc. RecSys. 11–18.
  40. Aleksandr V. Petrov and Craig Macdonald. 2022a. Effective and Efficient Training for Sequential Recommendation Using Recency Sampling. In Proc. RecSys. 81–91.
  41. Aleksandr V. Petrov and Craig Macdonald. 2022b. A Systematic Review and Replicability Study of BERT4Rec for Sequential Recommendation. In Proc. RecSys. 436–447.
  42. Aleksandr V. Petrov and Craig Macdonald. 2023. gSASRec: Reducing Overconfidence in Sequential Recommendation Trained with Negative Sampling. In Proc. RecSys. 116–128.
  43. Language Models are Unsupervised Multitask Learners. OpenAI blog (2019).
  44. Recommender Systems with Generative Retrieval. arXiv:2305.05065 [cs.IR]
  45. Steffen Rendle. 2022. Item Recommendation from Implicit Feedback. In Recommender Systems Handbook. 143–171.
  46. BPR: Bayesian Personalized Ranking from Implicit Feedback. In Proc. UAI.
  47. Revisiting the Performance of iALS on Item Recommendation Benchmarks. In Proc. RecSys. 427–435.
  48. ColBERTv2: Effective and Efficient Retrieval via Lightweight Late Interaction. arXiv:2112.01488 [cs]
  49. Ozan Sener and Vladlen Koltun. 2018. Multi-Task Learning as Multi-Objective Optimization. In Advances in Neural Information Processing Systems, Vol. 31.
  50. Compositional Embeddings Using Complementary Partitions for Memory-Efficient Recommendation Systems. In Proc. KDD. 165–175.
  51. Beyond User Embedding Matrix: Learning to Hash for Modeling Large-Scale Users in Recommendation. In Proc. SIGIR. 319–328.
  52. BERT4Rec: Sequential Recommendation with Bidirectional Encoder Representations from Transformer. In Proc. CIKM. 1441–1450.
  53. Jiaxi Tang and Ke Wang. 2018. Personalized Top-N Sequential Recommendation via Convolutional Sequence Embedding. In Proc. WSDM. 565–573.
  54. Attention is All you Need. In Proc. NeurIPS.
  55. Compressing Embedding Table via Multi-dimensional Quantization Encoding for Sequential Recommender Model. In Proc. ICCIP. 234–239.
  56. Deep & Cross Network for Ad Click Predictions. In Proc. KDD. 1–7.
  57. Feature Hashing for Large Scale Multitask Learning. arXiv:0902.2206 [cs]
  58. HuggingFace’s Transformers: State-of-the-art Natural Language Processing. arXiv:1910.03771 [cs]
  59. Efficient On-Device Session-Based Recommendation. ACM Transactions on Information Systems (2023).
  60. ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers. In Proc. NeurIPS, Vol. 35. 27168–27183.
  61. A Simple Cenvolutional Generative Network for Next Item Recommendation. In Proc. WSDM. 582–590.
  62. Calvin McCarter Fangrui Liu Hiroyuki Deguchi Lee Butterman Yusuke Matsui, Hiho Karuta. 2023. Nano Product Quantization (nanopq): a vanilla implementation of Product Quantization (PQ) and Optimized Product Quantization (OPQ) written in pure python without any third party dependencies. https://github.com/matsui528/nanopq. [Online; accessed 14-December-2023].
  63. Jointly Optimizing Query Encoder and Product Quantization to Improve Retrieval Performance. In Proc. CIKM. 2487–2496.
  64. Optimizing Dense Retrieval Model Training with Hard Negatives. In Proc. SIGIR. 1503–1512.
  65. A Model of Two Tales: Dual Transfer Learning Framework for Improved Long-tail Item Recommendation. In Proc. WWW. 2220–2231.
  66. A Personalized Recommendation Algorithm Based on Approximating the Singular Value Decomposition (ApproSVD). In Proc. WI-IAT, Vol. 2. 458–464.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Aleksandr V. Petrov (6 papers)
  2. Craig Macdonald (49 papers)
Citations (4)