Repeated Padding as Data Augmentation for Sequential Recommendation (2403.06372v1)
Abstract: Sequential recommendation aims to provide users with personalized suggestions based on their historical interactions. When training sequential models, padding is a widely adopted technique for two main reasons: 1) The vast majority of models can only handle fixed-length sequences; 2) Batching-based training needs to ensure that the sequences in each batch have the same length. The special value \emph{0} is usually used as the padding content, which does not contain the actual information and is ignored in the model calculations. This common-sense padding strategy leads us to a problem that has never been explored before: \emph{Can we fully utilize this idle input space by padding other content to further improve model performance and training efficiency?} In this paper, we propose a simple yet effective padding method called \textbf{Rep}eated \textbf{Pad}ding (\textbf{RepPad}). Specifically, we use the original interaction sequences as the padding content and fill it to the padding positions during model training. This operation can be performed a finite number of times or repeated until the input sequences' length reaches the maximum limit. Our RepPad can be viewed as a sequence-level data augmentation strategy. Unlike most existing works, our method contains no trainable parameters or hyperparameters and is a plug-and-play data augmentation operation. Extensive experiments on various categories of sequential models and five real-world datasets demonstrate the effectiveness and efficiency of our approach. The average recommendation performance improvement is up to 60.3\% on GRU4Rec and 24.3\% on SASRec. We also provide in-depth analysis and explanation of what makes RepPad effective from multiple perspectives. The source code will be released to ensure the reproducibility of our experiments.
- TiCoSeRec: Augmenting Data to Uniform Sequences by Time Intervals for Effective Recommendation. TKDE (2023).
- Uniform sequence better: Time interval aware data augmentation for sequential recommendation. In AAAI, Vol. 37. 4225–4232.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
- Lighter and better: low-rank decomposed self-attention networks for next-item recommendation. In SIGIR. 1733–1737.
- Sequential recommendation via stochastic self-attention. In WWW. 2036–2047.
- Learnable Model Augmentation Contrastive Learning for Sequential Recommendation. TKDE (2023).
- Ruining He and Julian McAuley. 2016. Fusing similarity models with markov chains for sparse sequential recommendation. In ICDM. IEEE, 191–200.
- Session-based recommendations with recurrent neural networks. arXiv preprint arXiv:1511.06939 (2015).
- Modeling personalized item frequency information for next-basket recommendation. In SIGIR. 1071–1080.
- Sequential recommendation with bidirectional chronological augmentation of transformer. arXiv preprint arXiv:2112.06460 (2021).
- Contrastive Self-supervised Learning in Recommender Systems: A Survey. arXiv preprint arXiv:2303.09902 (2023).
- Wang-Cheng Kang and Julian McAuley. 2018. Self-attentive sequential recommendation. In ICDM. IEEE, 197–206.
- Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
- Walid Krichene and Steffen Rendle. 2020. On sampled metrics for item recommendation. In KDD. 1748–1757.
- Neural attentive session-based recommendation. In CIKM. 1419–1428.
- Time interval aware self-attention for sequential recommendation. In WSDM. 322–330.
- Understanding the disharmony between dropout and batch normalization by variance shift. In CVPR. 2682–2690.
- Context-aware sequential recommendation. In ICDM. IEEE, 1053–1058.
- Diffusion augmentation for sequential recommendation. In CIKM. 1576–1586.
- Contrastive self-supervised sequential recommendation with robust augmentation. arXiv preprint arXiv:2108.06479 (2021).
- Augmenting sequential recommendation with pseudo-prior items via reversely pre-training transformer. In SIGIR. 1608–1612.
- Inferring networks of substitutable and complementary products. In KDD. 785–794.
- Intent Contrastive Learning with Cross Subsequences for Sequential Recommendation. arXiv preprint arXiv:2310.14318 (2023).
- Contrastive learning for representation degeneration problem in sequential recommendation. In WSDM. 813–823.
- CmnRec: Sequential Recommendations with Chunk-accelerated Memory Network. TKDE (2022).
- On the convergence of adam and beyond. arXiv preprint arXiv:1904.09237 (2019).
- Factorizing personalized markov chains for next-basket recommendation. In WWW. 811–820.
- Mike Schuster and Kuldip K Paliwal. 1997. Bidirectional recurrent neural networks. TSP 45, 11 (1997), 2673–2681.
- An MDP-based recommender system. JMLR 6, 9 (2005).
- BERT4Rec: Sequential recommendation with bidirectional encoder representations from transformer. In CIKM. 1441–1450.
- Improved recurrent neural networks for session-based recommendations. In Proceedings of the 1st workshop on deep learning for recommender systems. 17–22.
- Improved recurrent neural networks for session-based recommendations. In DLRS. 17–22.
- Jiaxi Tang and Ke Wang. 2018. Personalized top-n sequential recommendation via convolutional sequence embedding. In WSDM. 565–573.
- Periodicity May Be Emanative: Hierarchical Contrastive Learning for Sequential Recommendation. In CIKM. 2442–2451.
- Attention is all you need. NIPS 30 (2017).
- Learning to Augment for Casual User Recommendation. In WWW. 2183–2194.
- Counterfactual data-augmented sequential recommendation. In SIGIR. 347–356.
- Contrastive learning for sequential recommendation. arXiv preprint arXiv:2010.14395 (2020).
- Contrastive learning for sequential recommendation. In ICDE. IEEE, 1259–1273.
- Understanding and improving layer normalization. NIPS 32 (2019).
- Self-supervised learning for recommender systems: A survey. TKDE (2023).
- A simple convolutional generative network for next item recommendation. In WSDM. 582–590.
- S3-rec: Self-supervised learning for sequential recommendation with mutual information maximization. In CIKM. 1893–1902.
- Filter-enhanced MLP is all you need for sequential recommendation. In WWW. 2388–2399.
- Yizhou Dang (9 papers)
- Yuting Liu (62 papers)
- Enneng Yang (24 papers)
- Guibing Guo (35 papers)
- Linying Jiang (7 papers)
- Xingwei Wang (35 papers)
- Jianzhe Zhao (14 papers)