Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deep Natural Language Processing for LinkedIn Search (2108.13300v1)

Published 16 Aug 2021 in cs.IR and cs.CL

Abstract: Many search systems work with large amounts of natural language data, e.g., search queries, user profiles, and documents. Building a successful search system requires a thorough understanding of textual data semantics, where deep learning based natural language processing techniques (deep NLP) can be of great help. In this paper, we introduce a comprehensive study for applying deep NLP techniques to five representative tasks in search systems: query intent prediction (classification), query tagging (sequential tagging), document ranking (ranking), query auto completion (LLMing), and query suggestion (sequence to sequence). We also introduce BERT pre-training as a sixth task that can be applied to many of the other tasks. Through the model design and experiments of the six tasks, readers can find answers to four important questions: (1). When is deep NLP helpful/not helpful in search systems? (2). How to address latency challenges? (3). How to ensure model robustness? This work builds on existing efforts of LinkedIn search, and is tested at scale on LinkedIn's commercial search engines. We believe our experiences can provide useful insights for the industry and research communities.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Weiwei Guo (21 papers)
  2. Xiaowei Liu (137 papers)
  3. Sida Wang (21 papers)
  4. Michaeel Kazi (3 papers)
  5. Zhiwei Wang (223 papers)
  6. Zhoutong Fu (4 papers)
  7. Jun Jia (35 papers)
  8. Liang Zhang (357 papers)
  9. Huiji Gao (15 papers)
  10. Bo Long (60 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.