Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Selecting Query-bag as Pseudo Relevance Feedback for Information-seeking Conversations (2404.04272v1)

Published 22 Mar 2024 in cs.IR and cs.CL

Abstract: Information-seeking dialogue systems are widely used in e-commerce systems, with answers that must be tailored to fit the specific settings of the online system. Given the user query, the information-seeking dialogue systems first retrieve a subset of response candidates, then further select the best response from the candidate set through re-ranking. Current methods mainly retrieve response candidates based solely on the current query, however, incorporating similar questions could introduce more diverse content, potentially refining the representation and improving the matching process. Hence, in this paper, we proposed a Query-bag based Pseudo Relevance Feedback framework (QB-PRF), which constructs a query-bag with related queries to serve as pseudo signals to guide information-seeking conversations. Concretely, we first propose a Query-bag Selection module (QBS), which utilizes contrastive learning to train the selection of synonymous queries in an unsupervised manner by leveraging the representations learned from pre-trained VAE. Secondly, we come up with a Query-bag Fusion module (QBF) that fuses synonymous queries to enhance the semantic representation of the original query through multidimensional attention computation. We verify the effectiveness of the QB-PRF framework on two competitive pretrained backbone models, including BERT and GPT-2. Experimental results on two benchmark datasets show that our framework achieves superior performance over strong baselines.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. A multi domain knowledge enhanced matching network for response selection in retrieval-based dialogue systems. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 6597–6601. IEEE.
  2. Context-to-session matching: Utilizing whole session for response selection in information-seeking dialogue systems. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pages 1605–1613.
  3. Learning to respond with stickers: A framework of unifying multi-modality in multi-turn dialog. In Proceedings of the Web Conference 2020, pages 1138–1148.
  4. Learning to respond with your favorite stickers: A framework of unifying multi-modality and user preference in multi-turn dialog. ACM Transactions on Information Systems (TOIS), 39(2):1–32.
  5. Iseeq: Information seeking question generation using dynamic meta-information retrieval and knowledge graphs. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 10672–10680.
  6. Fine-grained post-training for improving retrieval-based dialogue systems. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 1549–1558.
  7. Diederik P Kingma and Max Welling. 2013. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114.
  8. Exploring dense retrieval for dialogue response selection. ACM Transactions on Information Systems, 42(3):1–29.
  9. Multi-intent attribute-aware text matching in searching. arXiv preprint arXiv:2402.07788.
  10. The ubuntu dialogue corpus: A large dataset for research in unstructured multi-turn dialogue systems. arXiv preprint arXiv:1506.08909.
  11. Triplenet: Triple attention network for multi-turn response selection in retrieval-based chatbots. arXiv preprint arXiv:1909.10666.
  12. Learning to expand: Reinforced response expansion for information-seeking conversations. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pages 4055–4064.
  13. Transfer learning for context-aware question matching in information-seeking conversations in e-commerce. arXiv preprint arXiv:1806.05434.
  14. Paul Solomon. 1997. Conversation in information-seeking contexts: A test of an analytical framework. Library & Information Science Research, 19(3):217–248.
  15. Qrfa: A data-driven model of information-seeking dialogues. In European conference on information retrieval, pages 541–557. Springer.
  16. Query resolution for conversational search with limited supervision. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, pages 921–930.
  17. Sequential matching network: A new architecture for multi-turn response selection in retrieval-based chatbots. arXiv preprint arXiv:1612.01627.
  18. Learning to respond with deep neural networks for retrieval-based human-computer conversation system. In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, pages 55–64.
  19. Iart: Intent-aware response ranking with transformers in information-seeking conversation systems. In Proceedings of The Web Conference 2020, pages 2592–2598.
  20. Response ranking with deep matching networks and external knowledge in information-seeking conversation systems. In The 41st international acm sigir conference on research & development in information retrieval, pages 245–254.
  21. Improving query representations for dense retrieval with pseudo relevance feedback. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pages 3592–3596.
  22. Multi-turn response selection for chatbots with deep attention matching network. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1118–1127.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Xiaoqing Zhang (30 papers)
  2. Xiuying Chen (80 papers)
  3. Shen Gao (49 papers)
  4. Shuqi Li (18 papers)
  5. Xin Gao (208 papers)
  6. Ji-Rong Wen (299 papers)
  7. Rui Yan (250 papers)
X Twitter Logo Streamline Icon: https://streamlinehq.com