Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DeText: A Deep Text Ranking Framework with BERT (2008.02460v1)

Published 6 Aug 2020 in cs.IR and cs.CL

Abstract: Ranking is the most important component in a search system. Mostsearch systems deal with large amounts of natural language data,hence an effective ranking system requires a deep understandingof text semantics. Recently, deep learning based natural languageprocessing (deep NLP) models have generated promising results onranking systems. BERT is one of the most successful models thatlearn contextual embedding, which has been applied to capturecomplex query-document relations for search ranking. However,this is generally done by exhaustively interacting each query wordwith each document word, which is inefficient for online servingin search product systems. In this paper, we investigate how tobuild an efficient BERT-based ranking model for industry use cases.The solution is further extended to a general ranking framework,DeText, that is open sourced and can be applied to various rankingproductions. Offline and online experiments of DeText on threereal-world search systems present significant improvement overstate-of-the-art approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Weiwei Guo (21 papers)
  2. Xiaowei Liu (137 papers)
  3. Sida Wang (21 papers)
  4. Huiji Gao (15 papers)
  5. Ananth Sankar (5 papers)
  6. Zimeng Yang (1 paper)
  7. Qi Guo (237 papers)
  8. Liang Zhang (357 papers)
  9. Bo Long (60 papers)
  10. Bee-Chung Chen (2 papers)
  11. Deepak Agarwal (7 papers)
Citations (22)

Summary

We haven't generated a summary for this paper yet.