Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Breaking the Hourglass Phenomenon of Residual Quantization: Enhancing the Upper Bound of Generative Retrieval (2407.21488v2)

Published 31 Jul 2024 in cs.IR and cs.AI

Abstract: Generative retrieval (GR) has emerged as a transformative paradigm in search and recommender systems, leveraging numeric-based identifier representations to enhance efficiency and generalization. Notably, methods like TIGER employing Residual Quantization-based Semantic Identifiers (RQ-SID), have shown significant promise in e-commerce scenarios by effectively managing item IDs. However, a critical issue termed the "\textbf{Hourglass}" phenomenon, occurs in RQ-SID, where intermediate codebook tokens become overly concentrated, hindering the full utilization of generative retrieval methods. This paper analyses and addresses this problem by identifying data sparsity and long-tailed distribution as the primary causes. Through comprehensive experiments and detailed ablation studies, we analyze the impact of these factors on codebook utilization and data distribution. Our findings reveal that the "Hourglass" phenomenon substantially impacts the performance of RQ-SID in generative retrieval. We propose effective solutions to mitigate this issue, thereby significantly enhancing the effectiveness of generative retrieval in real-world E-commerce applications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (14)
  1. Zhirui Kuai (1 paper)
  2. Zuxu Chen (2 papers)
  3. Huimu Wang (6 papers)
  4. Mingming Li (17 papers)
  5. Dadong Miao (5 papers)
  6. Binbin Wang (26 papers)
  7. Xusong Chen (1 paper)
  8. Li Kuang (8 papers)
  9. Yuxing Han (40 papers)
  10. Jiaxing Wang (16 papers)
  11. Guoyu Tang (12 papers)
  12. Lin Liu (190 papers)
  13. Songlin Wang (17 papers)
  14. Jingwei Zhuo (12 papers)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com