TF-DCon: Leveraging Large Language Models (LLMs) to Empower Training-Free Dataset Condensation for Content-Based Recommendation (2310.09874v4)
Abstract: Modern techniques in Content-based Recommendation (CBR) leverage item content information to provide personalized services to users, but suffer from resource-intensive training on large datasets. To address this issue, we explore the dataset condensation for textual CBR in this paper. The goal of dataset condensation is to synthesize a small yet informative dataset, upon which models can achieve performance comparable to those trained on large datasets. While existing condensation approaches are tailored to classification tasks for continuous data like images or embeddings, direct application of them to CBR has limitations. To bridge this gap, we investigate efficient dataset condensation for content-based recommendation. Inspired by the remarkable abilities of LLMs in text comprehension and generation, we leverage LLMs to empower the generation of textual content during condensation. To handle the interaction data involving both users and items, we devise a dual-level condensation method: content-level and user-level. At content-level, we utilize LLMs to condense all contents of an item into a new informative title. At user-level, we design a clustering-based synthesis module, where we first utilize LLMs to extract user interests. Then, the user interests and user embeddings are incorporated to condense users and generate interactions for condensed users. Notably, the condensation paradigm of this method is forward and free from iterative optimization on the synthesized dataset. Extensive empirical findings from our study, conducted on three authentic datasets, substantiate the efficacy of the proposed method. Particularly, we are able to approximate up to 97% of the original performance while reducing the dataset size by 95% (i.e., on dataset MIND).
- F. Wu, Y. Qiao, J.-H. Chen, C. Wu, T. Qi, J. Lian, D. Liu, X. Xie, J. Gao, W. Wu, and M. Zhou, “MIND: A large-scale dataset for news recommendation,” in Proc. of ACL’2020, 2020.
- C. Wu, F. Wu, M. An, J. Huang, Y. Huang, and X. Xie, “Neural news recommendation with attentive multi-view learning,” in Proc. of IJCAI’2019. ijcai.org, 2019.
- C. Wu, F. Wu, T. Qi, and Y. Huang, “Fastformer: Additive attention can be all you need,” arXiv preprint arXiv:2108.09084, 2021.
- H. Wang, F. Zhang, X. Xie, and M. Guo, “DKN: deep knowledge-aware network for news recommendation,” in In Proc. of WWW’2018. ACM, 2018.
- C. Wu, F. Wu, S. Ge, T. Qi, Y. Huang, and X. Xie, “Neural news recommendation with multi-head self-attention,” in In Proc. of EMNLP-IJCNLP’2019. ACL, 2019.
- Q. Liu, J. Zhu, J. Wu, T. Wu, Z. Dong, and X. Wu, “FANS: fast non-autoregressive sequence generation for item list continuation,” in Proc. of the Web Conference’2023. ACM, 2023.
- T. Wang, J.-Y. Zhu, A. Torralba, and A. A. Efros, “Dataset distillation,” arXiv preprint arXiv:1811.10959, 2018.
- W. Jin, L. Zhao, S. Zhang, Y. Liu, J. Tang, and N. Shah, “Graph condensation for graph neural networks,” in Proc. of ICLR’2022. OpenReview.net, 2022.
- X. Zheng, M. Zhang, C. Chen, Q. V. H. Nguyen, X. Zhu, and S. Pan, “Structure-free graph condensation: From large-scale graphs to condensed graph-free data,” CoRR, 2023.
- W. Jin, X. Tang, H. Jiang, Z. Li, D. Zhang, J. Tang, and B. Yin, “Condensing graphs via one-step gradient matching,” in Proc. of KDD’2022. ACM, 2022.
- B. Zhao, K. R. Mopuri, and H. Bilen, “Dataset condensation with gradient matching,” in Proc. of ICLR’2021. OpenReview.net, 2021.
- X. Han, A. Shen, Y. Li, L. Frermann, T. Baldwin, and T. Cohn, “Towards fair dataset distillation for text classification,” in In Proc. of SustaiNLP’2022. ACL, 2022.
- I. Sucholutsky and M. Schonlau, “Soft-label dataset distillation and text dataset distillation,” in 2021 International Joint Conference on Neural Networks (IJCNN). IEEE, 2021.
- N. Sachdeva, M. P. Dhaliwal, C. Wu, and J. J. McAuley, “Infinite recommendation networks: A data-centric approach,” in NeurIPS, 2022.
- J. Wu, W. Fan, S. Liu, Q. Liu, R. He, Q. Li, and K. Tang, “Dataset condensation for recommendation,” in arXiv, 2023.
- C. Wang, J. Sun, Z. Dong, R. Li, and R. Zhang, “Gradient matching for categorical data distillation in ctr prediction,” in Pro. of RecSys’2023. New York, NY, USA: ACM, 2023.
- T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell et al., “Language models are few-shot learners,” in In Proc. of NeurIPS’2020, 2020.
- A. Chowdhery, S. Narang, J. Devlin, M. Bosma, G. Mishra, A. Roberts, and et al, “Palm: Scaling language modeling with pathways,” 2022.
- J. Li, Y. Liu, W. Fan, X. Wei, H. Liu, J. Tang, and Q. Li, “Empowering molecule discovery for molecule-caption translation with large language models: A chatgpt perspective,” arXiv, 2023.
- F. Antaki, S. Touma, D. Milad, J. El-Khoury, and R. Duval, “Evaluating the performance of chatgpt in ophthalmology: An analysis of its successes and shortcomings.” in medRxiv, 2023.
- J. R. Benoit, “Chatgpt for clinical vignette generation, revision, and evaluation.” in medRxiv, 2023.
- W. Fan, Z. Zhao, J. Li, Y. Liu, X. Mei, Y. Wang, Z. Wen, F. Wang, X. Zhao, J. Tang, and Q. Li, “Recommender systems in the era of large language models (llms),” 2023.
- S. Dai, N. Shao, H. Zhao, W. Yu, Z. Si, C. Xu, Z. Sun, X. Zhang, and J. Xu, “Uncovering chatgpt’s capabilities in recommender systems,” in Proc. of RecSys’2023, 2023.
- Q. Liu, N. Chen, T. Sakai, and X.-M. Wu, “A first look at llm-powered generative news recommendation,” arXiv preprint arXiv:2305.06566, 2023.
- Z. Chen, H. Mao, H. Li, W. Jin, H. Wen, X. Wei, S. Wang, D. Yin, W. Fan, H. Liu, and J. Tang, “Exploring the potential of large language models (llms) in learning on graphs,” arXiv, 2023.
- H. Dai, Z. Liu, W. Liao, X. Huang, Y. Cao, Z. Wu, L. Zhao, S. Xu, W. Liu, N. Liu et al., “Auggpt: Leveraging chatgpt for text data augmentation,” arXiv preprint arXiv:2302.13007, 2023.
- S. Ubani, S. O. Polat, and R. Nielsen, “Zeroshotdataaug: Generating and augmenting training data with chatgpt,” arXiv preprint arXiv:2304.14334, 2023.
- J. Cui, R. Wang, S. Si, and C.-J. Hsieh, “Scaling up dataset distillation to ImageNet-1K with constant memory,” in Proc. of ICML’2023. PMLR, 2023.
- M. Wan and J. McAuley, “Item recommendation on monotonic behavior chains,” in Proc. of RecSys’2018, 2018.
- F. M. Harper and J. A. Konstan, “The movielens datasets: History and context,” ACM Trans. Interact. Intell. Syst., 2015.
- D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in In Proc. of ICLR’2015, 2015.
- L. Wu, X. He, X. Wang, K. Zhang, and M. Wang, “A survey on accuracy-oriented neural recommendation: From collaborative filtering to information-rich recommendation,” IEEE Trans. on Knowl. and Data Eng., 2022.
- A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in In Proc. of NeurIPS’2012, 2012, pp. 1106–1114.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, “Attention is all you need,” in In Proc. of NeurIPS’2017, 2017.
- J. Devlin, M. Chang, K. Lee, and K. Toutanova, “BERT: pre-training of deep bidirectional transformers for language understanding,” in In Proc. of NAACL-HLT’2019. ACL, 2019.
- J. Li, J. Zhu, Q. Bi, G. Cai, L. Shang, Z. Dong, X. Jiang, and Q. Liu, “MINER: multi-interest matching network for news recommendation,” in Findings of ACL’2022. ACL, 2022.
- Q. Zhang, J. Li, Q. Jia, C. Wang, J. Zhu, Z. Wang, and X. He, “UNBERT: user-news matching BERT for news recommendation,” in In Proc. of IJCAI’2021. ijcai.org, 2021.
- A. Maekawa, N. Kobayashi, K. Funakoshi, and M. Okumura, “Dataset distillation with attention labels for fine-tuning BERT,” in Proc. of ACL’2023. ACL, 2023.
- Y. Li and W. Li, “Data distillation for text classification,” in Proc. of SIGIR’2021. ACM, 2021.
- S. Sahni and H. Patel, “Exploring multilingual text data distillation,” 2023.
- X. Wu, Z. Deng, and O. Russakovsky, “Multimodal dataset distillation for image-text retrieval,” 2023.