Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Leave No One Behind: Online Self-Supervised Self-Distillation for Sequential Recommendation (2404.07219v2)

Published 22 Mar 2024 in cs.IR and cs.LG

Abstract: Sequential recommendation methods play a pivotal role in modern recommendation systems. A key challenge lies in accurately modeling user preferences in the face of data sparsity. To tackle this challenge, recent methods leverage contrastive learning (CL) to derive self-supervision signals by maximizing the mutual information of two augmented views of the original user behavior sequence. Despite their effectiveness, CL-based methods encounter a limitation in fully exploiting self-supervision signals for users with limited behavior data, as users with extensive behaviors naturally offer more information. To address this problem, we introduce a novel learning paradigm, named Online Self-Supervised Self-distillation for Sequential Recommendation ($S4$Rec), effectively bridging the gap between self-supervised learning and self-distillation methods. Specifically, we employ online clustering to proficiently group users by their distinct latent intents. Additionally, an adversarial learning strategy is utilized to ensure that the clustering procedure is not affected by the behavior length factor. Subsequently, we employ self-distillation to facilitate the transfer of knowledge from users with extensive behaviors (teachers) to users with limited behaviors (students). Experiments conducted on four real-world datasets validate the effectiveness of the proposed method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (39)
  1. Self-labelling via simultaneous clustering and representation learning. In ICLR.
  2. Unsupervised learning of visual features by contrasting cluster assignments. NeurIPS 33 (2020), 9912–9924.
  3. Wei-Chi Chen and Wei-Ta Chu. 2023. SSSD: Self-Supervised Self Distillation. In WACV. 2769–2776.
  4. Improving end-to-end sequential recommendations with intent-aware diversification. In CIKM. 175–184.
  5. Intent Contrastive Learning for Sequential Recommendation. In WWW. 2172–2182.
  6. Marco Cuturi. 2013. Sinkhorn distances: Lightspeed computation of optimal transport. NeurIPS 26 (2013).
  7. Yaroslav Ganin and Victor S. Lempitsky. 2015. Unsupervised Domain Adaptation by Backpropagation. In ICML, Vol. 37. 1180–1189.
  8. SimCSE: Simple Contrastive Learning of Sentence Embeddings. In EMNLP. 6894–6910.
  9. Generative Adversarial Nets. In NeurIPS. 2672–2680.
  10. Ruining He and Julian McAuley. 2016. Fusing similarity models with markov chains for sparse sequential recommendation. In ICDM. 191–200.
  11. Balázs Hidasi and Alexandros Karatzoglou. 2018. Recurrent neural networks with top-k gains for session-based recommendations. In CIKM. 843–852.
  12. Session-based recommendations with recurrent neural networks. ICLR (2015).
  13. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015).
  14. Improving sequential recommendation with knowledge-enhanced memory networks. In SIGIR. 505–514.
  15. Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In ICDM. IEEE, 197–206.
  16. Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
  17. Intention-aware Sequential Recommendation with Structured Intent Transition : (Extended Abstract). In ICDE. 3759–3760.
  18. Multi-Intention Oriented Contrastive Learning for Sequential Recommendation. In WSDM. 411–419.
  19. Dual Contrastive Network for Sequential Recommendation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR ’22). Association for Computing Machinery, New York, NY, USA, 2686–2691. https://doi.org/10.1145/3477495.3531918
  20. Contrastive Self-supervised Sequential Recommendation with Robust Augmentation. CoRR abs/2108.06479 (2021).
  21. Augmenting sequential recommendation with pseudo-prior items via reversely pre-training transformer. In SIGIR. 1608–1612.
  22. Disentangled Self-Supervision in Sequential Recommenders. In KDD. 483–491.
  23. Image-Based Recommendations on Styles and Substitutes. In SIGIR. 43–52.
  24. BPR: Bayesian Personalized Ranking from Implicit Feedback. In UAI. 452–461.
  25. Factorizing personalized markov chains for next-basket recommendation. In WWW. 811–820.
  26. BERT4Rec: Sequential recommendation with bidirectional encoder representations from transformer. In CIKM. 1441–1450.
  27. Sparse-interest network for sequential recommendation. In WSDM. 598–606.
  28. Jiaxi Tang and Ke Wang. 2018. Personalized top-n sequential recommendation via convolutional sequence embedding. In WSDM. 565–573.
  29. Representation Learning with Contrastive Predictive Coding. CoRR (2018). http://arxiv.org/abs/1807.03748
  30. Attention is All you Need. In NeurIPS. 5998–6008.
  31. Feng Wang and Huaping Liu. 2021. Understanding the Behaviour of Contrastive Loss. In CVPR. 2495–2504.
  32. Tongzhou Wang and Phillip Isola. 2020. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In ICML. 9929–9939.
  33. Self-Supervised Graph Co-Training for Session-based Recommendation. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management (CIKM ’21). Association for Computing Machinery, New York, NY, USA, 2180–2190. https://doi.org/10.1145/3459637.3482388
  34. Contrastive Learning for Sequential Recommendation. In ICDE. 1259–1273.
  35. Snapshot Distillation: Teacher-Student Optimization in One Generation. In CVPR. 2859–2868.
  36. Are graph augmentations necessary? simple graph contrastive learning for recommendation. In SIGIR. 1294–1303.
  37. A simple convolutional generative network for next item recommendation. In WSDM. 582–590.
  38. Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation. In ICCV. 3712–3721.
  39. S3-Rec: Self-Supervised Learning for Sequential Recommendation with Mutual Information Maximization. In CIKM. 1893–1902.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Shaowei Wei (3 papers)
  2. Zhengwei Wu (12 papers)
  3. Xin Li (980 papers)
  4. Qintong Wu (1 paper)
  5. Zhiqiang Zhang (129 papers)
  6. Jun Zhou (370 papers)
  7. Lihong Gu (8 papers)
  8. Jinjie Gu (50 papers)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com