2000 character limit reached
SPLADE-v3: New baselines for SPLADE (2403.06789v1)
Published 11 Mar 2024 in cs.IR and cs.CL
Abstract: A companion to the release of the latest version of the SPLADE library. We describe changes to the training structure and present our latest series of models -- SPLADE-v3. We compare this new version to BM25, SPLADE++, as well as re-rankers, and showcase its effectiveness via a meta-analysis over more than 40 query sets. SPLADE-v3 further pushes the limit of SPLADE models: it is statistically significantly more effective than both BM25 and SPLADE++, while comparing well to cross-encoder re-rankers. Specifically, it gets more than 40 MRR@10 on the MS MARCO dev set, and improves by 2% the out-of-domain results on the BEIR benchmark.
- E. Bassani. ranx: A blazing-fast python library for ranking evaluation and comparison. In European Conference on Information Retrieval, pages 259–264. Springer, 2022.
- Overview of the trec 2022 deep learning track. In Text Retrieval Conference, 2022.
- Benchmarking middle-trained language models for neural search. arXiv preprint arXiv:2306.02867, 2023.
- Splade v2: Sparse lexical and expansion model for information retrieval, 2021.
- From distillation to hard negative sampling: Making sparse neural ir models more effective. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2353–2359, 2022.
- Towards effective and efficient sparse neural information retrieval. ACM Trans. Inf. Syst., dec 2023. Just Accepted.
- SPLADE: Sparse Lexical and Expansion Model for First Stage Ranking. In Proc. SIGIR, page 2288–2292, 2021.
- L. Gao and J. Callan. Unsupervised corpus aware language model pre-training for dense passage retrieval. In S. Muresan, P. Nakov, and A. Villavicencio, editors, Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2843–2853, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Tevatron: An efficient and flexible toolkit for dense retrieval. arXiv preprint arXiv:2203.05765, 2022.
- Improving efficient neural ranking models with cross-architecture knowledge distillation, 2021.
- C. Lassance and S. Clinchant. An efficiency study for splade models. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2220–2226, 2022.
- C. Lassance and S. Clinchant. The tale of two ms marco – and their unfair comparisons, 2023.
- C. Lassance and S. Clinchant. The tale of two msmarco - and their unfair comparisons. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’23, page 2431–2435, New York, NY, USA, 2023. Association for Computing Machinery.
- Distilling dense representations for ranking using tightly-coupled teachers, 2020.
- Simplified data wrangling with ir_datasets. In SIGIR, 2021.
- Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter, 10 2019.
- ColBERTv2: Effective and efficient retrieval via lightweight late interaction. In M. Carpuat, M.-C. de Marneffe, and I. V. Meza Ruiz, editors, Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3715–3734, Seattle, United States, July 2022. Association for Computational Linguistics.
- Ranger: A toolkit for effect-size based multi-task evaluation. arXiv preprint arXiv:2305.15048, 2023.
- Exploring effect-size-based meta-analysis for multi-dataset evaluation. 2023.
- Beir: A heterogenous benchmark for zero-shot evaluation of information retrieval models. arXiv preprint arXiv:2104.08663, 2021.
- Curriculum learning for dense retrieval distillation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 1979–1983, 2022.
- Carlos Lassance (35 papers)
- Thibault Formal (17 papers)
- Stéphane Clinchant (39 papers)
- Hervé Déjean (16 papers)