Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Continuous Input Embedding Size Search For Recommender Systems (2304.03501v4)

Published 7 Apr 2023 in cs.IR

Abstract: Latent factor models are the most popular backbones for today's recommender systems owing to their prominent performance. Latent factor models represent users and items as real-valued embedding vectors for pairwise similarity computation, and all embeddings are traditionally restricted to a uniform size that is relatively large (e.g., 256-dimensional). With the exponentially expanding user base and item catalog in contemporary e-commerce, this design is admittedly becoming memory-inefficient. To facilitate lightweight recommendation, reinforcement learning (RL) has recently opened up opportunities for identifying varying embedding sizes for different users/items. However, challenged by search efficiency and learning an optimal RL policy, existing RL-based methods are restricted to highly discrete, predefined embedding size choices. This leads to a largely overlooked potential of introducing finer granularity into embedding sizes to obtain better recommendation effectiveness under a given memory budget. In this paper, we propose continuous input embedding size search (CIESS), a novel RL-based method that operates on a continuous search space with arbitrary embedding sizes to choose from. In CIESS, we further present an innovative random walk-based exploration strategy to allow the RL policy to efficiently explore more candidate embedding sizes and converge to a better decision. CIESS is also model-agnostic and hence generalizable to a variety of latent factor RSs, whilst experiments on two real-world datasets have shown state-of-the-art performance of CIESS under different memory budgets when paired with three popular recommendation models.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (54)
  1. Higher-Order Factorization Machines. In Proceedings of the 30th International Conference on Neural Information Processing Systems. 3359–3367.
  2. Try This Instead: Personalized and Interpretable Substitute Recommendation. In SIGIR. 891–900.
  3. Quaternion Factorization Machines: A Lightweight Solution to Intricate Feature Interaction Modeling. IEEE Transactions on Neural Networks and Learning Systems (2021), 1–14. https://doi.org/10.1109/TNNLS.2021.3118706
  4. Learning Elastic Embeddings for Customizing On-Device Recommenders. In SIGKDD. 138–147.
  5. Wide & deep learning for recommender systems. In Proceedings of the 1st workshop on deep learning for recommender systems. 7–10.
  6. Towards Automatic Discovering of Deep Hybrid Network Architecture for Sequential Recommendation. In WWW. 1923–1932.
  7. Deep reinforcement learning in large discrete action spaces. arXiv preprint arXiv:1512.07679 (2015).
  8. Addressing Function Approximation Error in Actor-Critic Methods. In Proceedings of the 35th International Conference on Machine Learning, Vol. 80. 1587–1596.
  9. Yuyun Gong and Qi Zhang. 2016. Hashtag recommendation using attention-based convolutional neural network.. In IJCAI. 2782–2788.
  10. A survey of actor-critic reinforcement learning: Standard and natural policy gradients. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 42, 6 (2012), 1291–1307.
  11. DeepFM: a factorization-machine based neural network for CTR prediction. IJCAI (2017).
  12. F Maxwell Harper and Joseph A Konstan. 2015. The movielens datasets: History and context. Acm transactions on interactive intelligent systems (tiis) 5, 4 (2015), 1–19.
  13. Lightgcn: Simplifying and powering graph convolution network for recommendation. In SIGIR. 639–648.
  14. Neural collaborative filtering. In WWW. 173–182.
  15. XLightFM: Extremely Memory-Efficient Factorization Machine. In SIGIR. 337–346.
  16. Neural input search for large scale recommendation models. In SIGKDD. 2387–2397.
  17. Learning to Embed Categorical Features without Embedding Tables for Recommendation. 840–850.
  18. Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In ICLR, Yoshua Bengio and Yann LeCun (Eds.).
  19. Walid Krichene and Steffen Rendle. 2022. On sampled metrics for item recommendation. Commun. ACM 65, 7 (2022), 75–83.
  20. Collaborative deep metric learning for video understanding. In Proceedings of the 24th ACM SIGKDD International conference on knowledge discovery & data mining. 481–490.
  21. Neural rating regression with abstractive tips generation for recommendation. In Proceedings of the 40th International ACM SIGIR conference on Research and Development in Information Retrieval. 345–354.
  22. Lightweight self-attentive sequential recommendation. In CIKM. 967–977.
  23. Lightrec: A memory and search-efficient recommender system. In WWW. 695–705.
  24. Continuous control with deep reinforcement learning. In ICLR.
  25. Automated embedding size search in deep recommender systems. In SIGIR. 2307–2316.
  26. Learnable Embedding sizes for Recommender Systems. In ICLR.
  27. Actor-Director-Critic: A Novel Deep Reinforcement Learning Framework. arXiv preprint arXiv:2301.03887 (2023).
  28. Autocross: Automatic feature crossing for tabular data in real-world applications. In SIGKDD. 1936–1945.
  29. OptEmbed: Learning Optimal Embedding Table for Click-through Rate Prediction. In CIKM. 1399–1409.
  30. A general method for automatic discovery of powerful interactions in click-through rate prediction. In SIGIR. 1298–1307.
  31. Single-Shot Embedding Dimension Search in Recommender System. In SIGIR. 513–522.
  32. Steffen Rendle. 2010. Factorization machines. In 2010 IEEE International conference on data mining. 995–1000.
  33. BPR: Bayesian Personalized Ranking from Implicit Feedback. In Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence. 452–461.
  34. Automatic selection of sparse matrix representation on GPUs. In ACM International Conference on Supercomputing. 99–108.
  35. Compositional embeddings using complementary partitions for memory-efficient recommendation systems. In SIGKDD. 165–175.
  36. Towards automated neural interaction discovery for click-through rate prediction. In SIGKDD. 945–955.
  37. Detecting beneficial feature interactions for recommender systems. In AAAI. 4357–4365.
  38. Detecting arbitrary order beneficial feature interactions for recommender systems. In SIGKDD. 1676–1686.
  39. G. E. Uhlenbeck and L. S. Ornstein. 1930. On the Theory of the Brownian Motion. Physical Review Journals 36 (1930), 823–841.
  40. Pauli Virtanen et al. 2020. SciPy 1.0: fundamental algorithms for scientific computing in Python. Nature Methods 17, 3 (2020), 261–272.
  41. Next Point-of-Interest Recommendation on Resource-Constrained Mobile Devices. In WWW. 906–916.
  42. A survey on session-based recommender systems. ACM Computing Surveys (CSUR) 54, 7 (2021), 1–38.
  43. What Your Images Reveal: Exploiting Visual Contents for Point-of-Interest Recommendation. In Proceedings of the 26th International Conference on World Wide Web. 391–400.
  44. Neural graph collaborative filtering. In SIGIR. 165–174.
  45. Autofield: Automating feature selection in deep recommender systems. In WWW. 1977–1986.
  46. On-Device Next-Item Recommendation with Self-Supervised Knowledge Distillation. In SIGIR. 546–555.
  47. Efficient On-Device Session-Based Recommendation. ACM Trans. Inf. Syst. 41, 4 (2023).
  48. Are Graph Augmentations Necessary? Simple Graph Contrastive Learning for Recommendation. In SIGIR. 1294–1303.
  49. Learn What Not to Learn: Action Elimination with Deep Reinforcement Learning. (2018), 3566–3577.
  50. Discrete Collaborative Filtering. In SIGIR. 325–334.
  51. Deep learning based recommender system: A survey and new perspectives. ACM Computing Surveys (CSUR) 52, 1 (2019), 1–38.
  52. AutoLoss: Automated Loss Function Search in Recommendations. In SIGKDD. 3959–3967.
  53. Autoemb: Automated embedding dimensionality search in streaming recommendations. In 2021 IEEE International Conference on Data Mining (ICDM). 896–905.
  54. AutoML for Deep Recommender Systems: A Survey. ACM Transactions on Information Systems (2022).
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Yunke Qu (4 papers)
  2. Tong Chen (200 papers)
  3. Xiangyu Zhao (192 papers)
  4. Lizhen Cui (66 papers)
  5. Kai Zheng (134 papers)
  6. Hongzhi Yin (210 papers)
Citations (12)