TPRF: A Transformer-based Pseudo-Relevance Feedback Model for Efficient and Effective Retrieval (2401.13509v2)
Abstract: This paper considers Pseudo-Relevance Feedback (PRF) methods for dense retrievers in a resource constrained environment such as that of cheap cloud instances or embedded systems (e.g., smartphones and smartwatches), where memory and CPU are limited and GPUs are not present. For this, we propose a transformer-based PRF method (TPRF), which has a much smaller memory footprint and faster inference time compared to other deep LLMs that employ PRF mechanisms, with a marginal effectiveness loss. TPRF learns how to effectively combine the relevance feedback signals from dense passage representations. Specifically, TPRF provides a mechanism for modelling relationships and weights between the query and the relevance feedback signals. The method is agnostic to the specific dense representation used and thus can be generally applied to any dense retriever.
- Hiteshwar Kumar Azad and Akshay Deepak. 2019. Query expansion techniques for information retrieval: a survey. Information Processing & Management 56, 5 (2019), 1698–1735.
- Overview of the TREC 2019 Deep Learning Track. In Text REtrieval Conference, TREC.
- Overview of the TREC 2020 Deep Learning Track. In Text REtrieval Conference, TREC.
- Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2978–2988.
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 4171–4186.
- A deep look into neural ranking models for information retrieval. Information Processing & Management 57, 6 (2020), 102067.
- Improving efficient neural ranking models with cross-architecture knowledge distillation. arXiv preprint arXiv:2010.02666 (2020).
- Efficiently Teaching an Effective Dense Retriever with Balanced Topic Aware Sampling. arXiv preprint arXiv:2104.06967 (2021).
- Billion-scale similarity search with gpus. IEEE Transactions on Big Data (2019).
- Pseudo relevance feedback with deep language models and dense retrievers: Successes and pitfalls. arXiv preprint arXiv:2108.11044 (2021).
- Improving Query Representations for Dense Retrieval with Pseudo Relevance Feedback: A Reproducibility Study. In Proceedings of the 44rd European Conference on Information Retrieval.
- Distilling dense representations for ranking using tightly-coupled teachers. arXiv preprint arXiv:2010.11386 (2020).
- In-batch negatives for knowledge distillation with tightly-coupled teachers for dense retrieval. In Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021). 163–173.
- MS MARCO: A Human Generated Machine Reading Comprehension Dataset. In Workshop on Cognitive Computing at NIPS.
- Rethinking Query Expansion for BERT Reranking. In Advances in Information Retrieval-42nd European Conference on IR Research, ECIR 2020.
- RocketQA: An Optimized Training Approach to Dense Passage Retrieval for Open-Domain Question Answering. arXiv preprint arXiv:2010.08191 (2020).
- Language Models Are Unsupervised Multitask Learners. In OpenAI blog.
- Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research 21 (2020), 1–67.
- Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics.
- RocketQAv2: A Joint Training Method for Dense Passage Retrieval and Passage Re-ranking. arXiv preprint arXiv:2110.07367 (2021).
- Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems. 6000–6010.
- A Pseudo-Relevance Feedback Framework Combining Relevance Matching and Semantic Matching for Information Retrieval. Information Processing & Management 57, 6 (2020), 102342.
- Pseudo-Relevance Feedback for Multiple Representation Dense Retrieval. arXiv preprint arXiv:2106.11251 (2021).
- Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. 38–45.
- Approximate Nearest Neighbor Negative Contrastive Learning For Dense Text Retrieval. arXiv preprint arXiv:2007.00808 (2020).
- XLNet: Generalized Autoregressive Pretraining for Language Understanding. Advances in Neural Information Processing Systems 32 (2019), 5753–5763.
- Pretrained Transformers for Text Ranking: BERT and Beyond. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining. 1154–1156.
- PGT: Pseudo Relevance Feedback Using a Graph-Based Transformer. In European Conference on Information Retrieval.
- Improving Query Representations for Dense Retrieval with Pseudo Relevance Feedback. In Proceedings of the 30th ACM International Conference on Information and Knowledge Management.
- RepBERT: Contextualized Text Embeddings for First-Stage Retrieval. arXiv preprint arXiv:2006.15498 (2020).
- BERT-QE: Contextualized Query Expansion for Document Re-ranking. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings. 4718–4728.
- Contextualized query expansion via unsupervised chunk selection for text retrieval. Information Processing & Management 58, 5 (2021), 102672.
- Chuting Yu (1 paper)
- Hang Li (277 papers)
- Ahmed Mourad (8 papers)
- Bevan Koopman (37 papers)
- Guido Zuccon (73 papers)