Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-task Sentence Encoding Model for Semantic Retrieval in Question Answering Systems (1911.07405v1)

Published 18 Nov 2019 in cs.CL, cs.AI, and cs.IR

Abstract: Question Answering (QA) systems are used to provide proper responses to users' questions automatically. Sentence matching is an essential task in the QA systems and is usually reformulated as a Paraphrase Identification (PI) problem. Given a question, the aim of the task is to find the most similar question from a QA knowledge base. In this paper, we propose a Multi-task Sentence Encoding Model (MSEM) for the PI problem, wherein a connected graph is employed to depict the relation between sentences, and a multi-task learning model is applied to address both the sentence matching and sentence intent classification problem. In addition, we implement a general semantic retrieval framework that combines our proposed model and the Approximate Nearest Neighbor (ANN) technology, which enables us to find the most similar question from all available candidates very quickly during online serving. The experiments show the superiority of our proposed method as compared with the existing sentence matching models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Qiang Huang (50 papers)
  2. Jianhui Bu (1 paper)
  3. Weijian Xie (10 papers)
  4. Shengwen Yang (5 papers)
  5. Weijia Wu (47 papers)
  6. Liping Liu (26 papers)
Citations (16)

Summary

We haven't generated a summary for this paper yet.