Hi-Gen: Generative Retrieval For Large-Scale Personalized E-commerce Search (2404.15675v2)
Abstract: Leveraging generative retrieval (GR) techniques to enhance search systems is an emerging methodology that has shown promising results in recent years. In GR, a text-to-text model maps string queries directly to relevant document identifiers (docIDs), dramatically simplifying the retrieval process. However, when applying most GR models in large-scale E-commerce for personalized item search, we must face two key problems in encoding and decoding. (1) Existing docID generation methods ignore the encoding of efficiency information, which is critical in E-commerce. (2) The positional information is important in decoding docIDs, while prior studies have not adequately discriminated the significance of positional information or well exploited the inherent interrelation among these positions. To overcome these problems, we introduce an efficient Hierarchical encoding-decoding Generative retrieval method (Hi-Gen) for large-scale personalized E-commerce search systems. Specifically, we first design a representation learning model using metric learning to learn discriminative feature representations of items to capture semantic relevance and efficiency information. Then, we propose a category-guided hierarchical clustering scheme that makes full use of the semantic and efficiency information of items to facilitate docID generation. Finally, we design a position-aware loss to discriminate the importance of positions and mine the inherent interrelation between different tokens at the same position. This loss boosts the performance of the LLM used in the decoding stage. Besides, we propose two variants of Hi-Gen (Hi-Gen-I2I and Hi-Gen-Cluster) to support online real-time large-scale recall in the online serving process. Hi-Gen gets 3.30% and 4.62% improvements over SOTA for Recall@1 on the public and industry datasets, respectively.
- Autoregressive search engines: Generating substrings as document identifiers. Advances in Neural Information Processing Systems 35 (2022), 31668–31683.
- Sergey Brin and Lawrence Page. 1998. The anatomy of a large-scale hypertextual Web search engine. In Proceedings of the Seventh International Conference on World Wide Web 7. Elsevier Science Publishers B. V., NLD, 107–117.
- Zhuyun Dai and Jamie Callan. 2019. Context-aware sentence/passage term importance estimation for first stage retrieval. arXiv preprint arXiv:1910.10687 (2019).
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
- Challenges and Advances in Generative Information Retrieval. In Proceedings of the 22nd Chinese National Conference on Computational Linguistics (Volume 2: Frontier Forum). 57–66.
- AOL4PS: A Large-Scale Dataset for Personalized Search. Data Intelligence 3 (08 2021), 1–17. https://doi.org/10.1162/dint_a_00104
- Elad Hoffer and Nir Ailon. 2015. Deep metric learning using triplet network. In Similarity-Based Pattern Recognition: Third International Workshop, SIMBAD 2015, Copenhagen, Denmark, October 12-14, 2015. Proceedings 3. Springer, 84–92.
- How to Index Item IDs for Recommendation Foundation Models. arXiv preprint arXiv:2305.06569 (2023).
- Dense passage retrieval for open-domain question answering. arXiv preprint arXiv:2004.04906 (2020).
- Mahmut Kaya and Hasan Şakir Bilge. 2019. Deep metric learning: A survey. Symmetry 11, 9 (2019), 1066.
- IncDSI: incrementally updatable document retrieval. In International Conference on Machine Learning. PMLR, 17122–17134.
- Brian Kulis et al. 2013. Metric learning: A survey. Foundations and Trends® in Machine Learning 5, 4 (2013), 287–364.
- Generative retrieval for long sequences. arXiv preprint arXiv:2204.13596 (2022).
- Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019).
- PersonalTM: Transformer memory for personalized retrieval. (2023).
- Amazon.com recommendations: item-to-item collaborative filtering. IEEE Internet Computing 7, 1 (2003), 76–80. https://doi.org/10.1109/MIC.2003.1167344
- Amazon.com Recommendations: Item-to-Item Collaborative Filtering. IEEE Internet Computing 7 (2003), 76–80.
- RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv:1907.11692 [cs.CL]
- DSI++: Updating Transformer Memory with New Documents. arXiv preprint arXiv:2212.09744 (2022).
- A Fast and Flexible Algorithm for Solving the Lasso in Large-scale and Ultrahigh-dimensional Problems. bioRxiv (2019). https://doi.org/10.1101/630079
- Language Models are Unsupervised Multitask Learners. (2019).
- Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research 21, 1 (2020), 5485–5551.
- Recommender Systems with Generative Retrieval. arXiv preprint arXiv:2305.05065 (2023).
- LaMP: When Large Language Models Meet Personalization. arXiv preprint arXiv:2304.11406 (2023).
- Learning to Tokenize for Generative Retrieval. arXiv preprint arXiv:2304.04171 (2023).
- Semantic-Enhanced Differentiable Search Index Inspired by Learning Strategies. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (, Long Beach, CA, USA,) (KDD ’23). Association for Computing Machinery, 4904–4913.
- Transformer memory as a differentiable search index. Advances in Neural Information Processing Systems 35 (2022), 21831–21843.
- Billion-scale commodity embedding for e-commerce recommendation in alibaba. In Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining. 839–848.
- A neural corpus indexer for document retrieval. Advances in Neural Information Processing Systems 35 (2022), 25600–25614.
- Liu Yang and Rong Jin. 2006. Distance metric learning: A comprehensive survey. Michigan State Universiy 2, 2 (2006), 4.
- Large Scale Product Graph Construction for Recommendation in E-commerce. ArXiv abs/2010.05525 (2020). https://api.semanticscholar.org/CorpusID:222291351
- Deepspeed-chat: Easy, fast and affordable rlhf training of chatgpt-like models at all scales. arXiv preprint arXiv:2308.01320 (2023).
- Generate rather than retrieve: Large language models are strong context generators. arXiv preprint arXiv:2209.10063 (2022).
- A dual augmented two-tower model for online large-scale recommendation. DLP-KDD (2021).
- A Personalized Dense Retrieval Framework for Unified Information Access.
- Towards personalized and semantic retrieval: An end-to-end solution for e-commerce search via embedding learning. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval. 2407–2416.
- Dense text retrieval based on pretrained language models: A survey. arXiv preprint arXiv:2211.14876 (2022).
- Guoqing Zheng and Jamie Callan. 2015. Learning to reweight terms with distributed representations. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval. 575–584.
- Encoding history with context-aware representation learning for personalized search. In Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval. 1111–1120.
- Ultron: An ultimate retriever on corpus with a model-based indexer. arXiv preprint arXiv:2208.09257 (2022).
- DynamicRetriever: A Pre-trained Model-based IR System Without an Explicit Index. Machine Intelligence Research 20, 2 (2023), 276–288.
- Bridging the gap between indexing and retrieval for differentiable search index with query generation. arXiv preprint arXiv:2206.10128 (2022).
- Yanjing Wu (1 paper)
- Yinfu Feng (2 papers)
- Jian Wang (967 papers)
- Wenji Zhou (1 paper)
- Yunan Ye (3 papers)
- Rong Xiao (44 papers)
- Jun Xiao (134 papers)