Enhancing Scalability in Recommender Systems through Lottery Ticket Hypothesis and Knowledge Distillation-based Neural Network Pruning (2401.10484v1)
Abstract: This study introduces an innovative approach aimed at the efficient pruning of neural networks, with a particular focus on their deployment on edge devices. Our method involves the integration of the Lottery Ticket Hypothesis (LTH) with the Knowledge Distillation (KD) framework, resulting in the formulation of three distinct pruning models. These models have been developed to address scalability issue in recommender systems, whereby the complexities of deep learning models have hindered their practical deployment. With judicious application of the pruning techniques, we effectively curtail the power consumption and model dimensions without compromising on accuracy. Empirical evaluation has been performed using two real world datasets from diverse domains against two baselines. Gratifyingly, our approaches yielded a GPU computation-power reduction of up to 66.67%. Notably, our study contributes to the field of recommendation system by pioneering the application of LTH and KD.
- Abarja, R. 2020. “Movie Rating Prediction using Convolutional Neural Network based on Historical Values,” International Journal of Emerging Trends in Engineering Research (8), pp. 2156–2164.
- Alford, S., Robinett, R., Milechin, L., and Kepner, J. 2018. “Pruned and Structurally Sparse Neural Networks,” In: 2018 IEEE MIT Undergraduate Research Technology Conference (URTC).
- Asami, T., Masumura, R., Yamaguchi, Y., Masataki, H., and Aono, Y. 2017. “Domain adaptation of DNN acoustic models using knowledge distillation,” In: 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
- Bharadhwaj, M., Ramadurai, G., and Ravindran, B. 2022. “Detecting Vehicles on the Edge: Knowledge Distillation to Improve Performance in Heterogeneous Road Traffic,” In: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
- Chen, H., Fu, J., Zhang, L., Wang, S., Lin, K., Shi, L., and Wang, L. 2019. “Deformable Convolutional Matrix Factorization for Document Context-Aware Recommendation in Social Networks,” IEEE Access (7), pp. 66,347–66,357.
- Frankle, J., and Carbin, M. 2019. “The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks,” .
- Girish, S., Maiya, S. R., Gupta, K., Chen, H., Davis, L. S., and Shrivastava, A. 2021. “The Lottery Ticket Hypothesis for Object Recognition,” In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
- Gordon, A., Eban, E., Nachum, O., Chen, B., Wu, H., Yang, T.-J., and Choi, E. 2018. “MorphNet: Fast & Simple Resource-Constrained Structure Learning of Deep Networks,” .
- Herlocker, J. L., Konstan, J. A., Terveen, L. G., and Riedl, J. T. 2004. “Evaluating Collaborative Filtering Recommender Systems,” ACM Trans Inf Syst (22:1), p. 5–53. URL https://doi.org/10.1145/963770.963772
- Hinton, G., Vinyals, O., and Dean, J. 2015. “Distilling the Knowledge in a Neural Network,” .
- Hou, S., Pan, X., Loy, C. C., Wang, Z., and Lin, D. 2018. “Lifelong Learning via Progressive Distillation and Retrospection,” In: Ferrari, V., Hebert, M., Sminchisescu, C., and Weiss, Y. (eds.) Computer Vision – ECCV 2018, Cham: Springer International Publishing.
- Ji, M., Heo, B., and Park, S. 2021. “Show, Attend and Distill:Knowledge Distillation via Attention-based Feature Matching,” .
- Krizhevsky, A. 2012. “Learning Multiple Layers of Features from Tiny Images,” University of Toronto .
- Kumar, P., and Thakur, R. S. 2018. “Recommendation system techniques and related issues: a survey,” International Journal of Information Technology (10), pp. 495–501.
- Li, Z., and Hoiem, D. 2017. “Learning without Forgetting,” .
- Molchanov, P., Tyree, S., Karras, T., Aila, T., and Kautz, J. 2017. “Pruning Convolutional Neural Networks for Resource Efficient Inference,” .
- Nagarnaik, P., and Thomas, A. 2015. “Survey on recommendation system methods,” In: 2015 2nd International Conference on Electronics and Communication Systems (ICECS).
- Nasraoui, O., and Petenes, C. 2003. “An intelligent Web recommendation engine based on fuzzy approximate reasoning,” In: The 12th IEEE International Conference on Fuzzy Systems, 2003. FUZZ ’03., vol. 2.
- Orbes-Arteaga, M., Cardoso, J., Sørensen, L., Igel, C., Ourselin, S., Modat, M., Nielsen, M., and Pai, A. 2019. “Knowledge distillation for semi-supervised domain adaptation,” .
- Romero, A., Ballas, N., Kahou, S. E., Chassang, A., Gatta, C., and Bengio, Y. 2015. “FitNets: Hints for Thin Deep Nets,” .
- Tian, Y., Krishnan, D., and Isola, P. 2022. “Contrastive Representation Distillation,” .
- Xia, M., Zhong, Z., and Chen, D. 2022. “Structured Pruning Learns Compact and Accurate Models,” .
- Xie, H., Jiang, W., Luo, H., and Yu, H. 2021. “Model compression via pruning and knowledge distillation for person re-identification,” Journal of Ambient Intelligence and Humanized Computing (12).
- Xu, K., Ba, J., Kiros, R., Cho, K., Courville, A., Salakhutdinov, R., Zemel, R., and Bengio, Y. 2016. “Show, Attend and Tell: Neural Image Caption Generation with Visual Attention,” .
- Yim, J., Joo, D., Bae, J., and Kim, J. 2017. “A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning,” In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
- Zagoruyko, S., and Komodakis, N. 2016. “Wide Residual Networks,” .
- Zhang, S., Yao, L., and Sun, A. 2017. “Deep Learning based Recommender System: A Survey and New Perspectives,” CoRR (abs/1707.07435). URL http://arxiv.org/abs/1707.07435
- Zhou, B., Hui, S., and Chang, K. 2004. “An intelligent recommender system using sequential Web access patterns,” In: IEEE Conference on Cybernetics and Intelligent Systems, 2004., vol. 1.
- Zhou, G., Mou, N., Fan, Y., Pi, Q., Bian, W., Zhou, C., Zhu, X., and Gai, K. 2018. “Deep Interest Evolution Network for Click-Through Rate Prediction,” .
- Rajaram R (1 paper)
- Manoj Bharadhwaj (1 paper)
- Vasan VS (1 paper)
- Nargis Pervin (1 paper)