Papers
Topics
Authors
Recent
2000 character limit reached

SimRec: Mitigating the Cold-Start Problem in Sequential Recommendation by Integrating Item Similarity (2410.22136v1)

Published 29 Oct 2024 in cs.IR

Abstract: Sequential recommendation systems often struggle to make predictions or take action when dealing with cold-start items that have limited amount of interactions. In this work, we propose SimRec - a new approach to mitigate the cold-start problem in sequential recommendation systems. SimRec addresses this challenge by leveraging the inherent similarity among items, incorporating item similarities into the training process through a customized loss function. Importantly, this enhancement is attained with identical model architecture and the same amount of trainable parameters, resulting in the same inference time and requiring minimal additional effort. This novel approach results in a robust contextual sequential recommendation model capable of effectively handling rare items, including those that were not explicitly seen during training, thereby enhancing overall recommendation performance. Rigorous evaluations against multiple baselines on diverse datasets showcase SimRec's superiority, particularly in scenarios involving items occurring less than 10 times in the training data. The experiments reveal an impressive improvement, with SimRec achieving up to 78% higher HR@10 compared to SASRec. Notably, SimRec outperforms strong baselines on sparse datasets while delivering on-par performance on dense datasets. Our code is available at https://github.com/amazon-science/sequential-recommendation-using-similarity.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. Dcdir: A deep cross-domain recommendation system for cold start users in insurance domain. In Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval, pages 1661–1664, 2020.
  2. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  3. Addressing cold start for next-song recommendation. In Proceedings of the 10th ACM Conference on Recommender Systems, pages 115–118, 2016.
  4. Movie genome: alleviating new item cold start in movie recommendation. User Modeling and User-Adapted Interaction, 29:291–343, 2019.
  5. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  6. X. Glorot and Y. Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pages 249–256. JMLR Workshop and Conference Proceedings, 2010.
  7. J. Gope and S. K. Jain. A survey on solving cold start problem in recommender systems. In 2017 International Conference on Computing, Communication and Automation (ICCCA), pages 133–138, 2017. doi: 10.1109/CCAA.2017.8229786.
  8. The movielens datasets: History and context. Acm transactions on interactive intelligent systems (tiis), 5(4):1–19, 2015.
  9. R. He and J. McAuley. Ups and downs: Modeling the visual evolution of fashion trends with one-class collaborative filtering. In proceedings of the 25th international conference on world wide web, pages 507–517, 2016.
  10. Session-based recommendations with recurrent neural networks. arXiv preprint arXiv:1511.06939, 2015.
  11. Learning to learn a cold-start sequential recommender. ACM Transactions on Information Systems (TOIS), 40(2):1–25, 2022.
  12. W.-C. Kang and J. McAuley. Self-attentive sequential recommendation. In 2018 IEEE international conference on data mining (ICDM), pages 197–206. IEEE, 2018.
  13. D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  14. Reformer: The efficient transformer. arXiv preprint arXiv:2001.04451, 2020.
  15. Time interval aware self-attention for sequential recommendation. In Proceedings of the 13th international conference on web search and data mining, pages 322–330, 2020.
  16. Text is all you need: Learning language representations for sequential recommendation. arXiv preprint arXiv:2305.13731, 2023a.
  17. Towards general text embeddings with multi-stage contrastive learning. arXiv preprint arXiv:2308.03281, 2023b.
  18. Image-based recommendations on styles and substitutes. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval, pages 43–52, 2015.
  19. Sentence-t5: Scalable sentence encoders from pre-trained text-to-text models. arXiv preprint arXiv:2108.08877, 2021.
  20. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  21. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485–5551, 2020.
  22. Context and attribute-aware sequential recommendation via cross-attention. In Proceedings of the 16th ACM Conference on Recommender Systems, pages 71–80, 2022.
  23. N. Reimers and I. Gurevych. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084, 2019.
  24. N. Reimers and I. Gurevych. The curse of dense low-dimensional information retrieval for large index sizes. arXiv preprint arXiv:2012.14210, 2020.
  25. Financial news recommendation based on graph embeddings. Decision Support Systems, 125:113115, 2019.
  26. Bert4rec: Sequential recommendation with bidirectional encoder representations from transformer. In Proceedings of the 28th ACM international conference on information and knowledge management, pages 1441–1450, 2019.
  27. Survey on collaborative and content-based recommendation systems. In 2023 5th International Conference on Smart Systems and Inventive Technology (ICSSIT), pages 1541–1548, 2023. doi: 10.1109/ICSSIT55814.2023.10061072.
  28. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  29. Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers. Advances in Neural Information Processing Systems, 33:5776–5788, 2020.
  30. Where to go next for recommender systems? id-vs. modality-based recommender models revisited. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2639–2649, 2023.
  31. Cold-start sequential recommendation via meta learner. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 4706–4713, 2021.
  32. S3-rec: Self-supervised learning for sequential recommendation with mutual information maximization. In Proceedings of the 29th ACM international conference on information & knowledge management, pages 1893–1902, 2020.

Summary

  • The paper presents SimRec, a method that integrates item similarity into sequential recommendation frameworks to mitigate the cold-start problem.
  • It employs a novel loss function that combines binary cross-entropy with a custom similarity loss derived from text embeddings.
  • Experiments on Amazon reviews and benchmark datasets demonstrate up to 78% HR@10 improvement, proving its effectiveness on sparse data.

SimRec: Mitigating the Cold-Start Problem in Sequential Recommendation Systems

The paper "SimRec: Mitigating the Cold-Start Problem in Sequential Recommendation by Integrating Item Similarity," authored by Shaked Brody and Shoval Lagziel, explores an innovative approach to address the cold-start problem in sequential recommendation systems. This issue, prevalent in domains such as e-commerce and content streaming, emerges when systems encounter items with few or no interactions, impeding accurate predictions.

Methodology and Technical Contribution

The core contribution of this research is SimRec, an approach that leverages item similarity to enhance the recommendation process. Unlike many existing methods, SimRec introduces only minimal changes to the model architecture and does not increase inference time, making it highly efficient. The researchers employ a novel loss function, integrating similarity information into the training process without extending learnable parameters.

The methodology encompasses computing item similarities, primarily through text embeddings derived via a pretrained LLM. Cosine similarity from these embeddings helps form a distribution reflecting item similarity across the dataset. Importantly, the proposed loss function, LSimRec\mathcal{L}_{SimRec}, merges binary cross-entropy loss with a custom similarity loss. This latter component compares the model’s prediction distribution against an item's similarity distribution, thereby refining the model's understanding of cold-start items.

Experimental Evaluation

The authors rigorously evaluate SimRec across both sparse and dense datasets. These experiments were performed on Amazon reviews datasets and well-known benchmarks such as ML-1M and Steam. SimRec consistently demonstrates superior performance on datasets characterized by sparsity, notably improving HR@10 by up to 78% over SASRec for items appearing fewer than 10 times in the training set.

A breakdown of performance against item frequency highlights SimRec's strength in predicting cold-start items, where it substantially outweighs SASRec. For datasets where rare items prevail, SimRec's enhancements in recommendation accuracy are evident and significant.

Theoretical and Practical Implications

From a theoretical standpoint, the research substantiates the hypothesis that incorporating item similarity into the learning process can significantly mitigate the cold-start problem. By presenting a method agnostic to specific dataset characteristics or domain-specific knowledge beyond textual features, SimRec sets a precedent for embedding similarity-based approaches into diverse recommendation systems.

Practically, SimRec’s advantage lies in its compatibility with existing architectures and datasets, requiring minimal computational overhead while delivering robust performance improvements. This is particularly beneficial in real-world applications where computational resources and inference latency are critical factors.

Future Directions

The study opens avenues for future exploration. One potential direction could involve assessing the integration of other data modalities beyond text, enhancing the similarity calculation phase with multimodal embeddings. Additionally, further analysis of the scalability of SimRec with expanding dataset sizes could provide insights into its utility in even broader applications.

In conclusion, SimRec represents a significant stride in managing cold-start challenges within sequential recommendation environments. By harmoniously integrating item similarity into the recommendation framework, it offers a promising solution with practical relevance and lays groundwork for continued advancements in this domain.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 15 likes about this paper.