Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Survey for Efficient Open Domain Question Answering (2211.07886v1)

Published 15 Nov 2022 in cs.CL

Abstract: Open domain question answering (ODQA) is a longstanding task aimed at answering factual questions from a large knowledge corpus without any explicit evidence in NLP. Recent works have predominantly focused on improving the answering accuracy and achieved promising progress. However, higher accuracy often comes with more memory consumption and inference latency, which might not necessarily be efficient enough for direct deployment in the real world. Thus, a trade-off between accuracy, memory consumption and processing speed is pursued. In this paper, we provide a survey of recent advances in the efficiency of ODQA models. We walk through the ODQA models and conclude the core techniques on efficiency. Quantitative analysis on memory cost, processing speed, accuracy and overall comparison are given. We hope that this work would keep interested scholars informed of the advances and open challenges in ODQA efficiency research, and thus contribute to the further development of ODQA efficiency.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Qin Zhang (98 papers)
  2. Shangsi Chen (1 paper)
  3. Dongkuan Xu (43 papers)
  4. Qingqing Cao (16 papers)
  5. Xiaojun Chen (100 papers)
  6. Trevor Cohn (105 papers)
  7. Meng Fang (100 papers)
Citations (29)