Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

K-AID: Enhancing Pre-trained Language Models with Domain Knowledge for Question Answering (2109.10547v1)

Published 22 Sep 2021 in cs.AI

Abstract: Knowledge enhanced pre-trained LLMs (K-PLMs) are shown to be effective for many public tasks in the literature but few of them have been successfully applied in practice. To address this problem, we propose K-AID, a systematic approach that includes a low-cost knowledge acquisition process for acquiring domain knowledge, an effective knowledge infusion module for improving model performance, and a knowledge distillation component for reducing the model size and deploying K-PLMs on resource-restricted devices (e.g., CPU) for real-world application. Importantly, instead of capturing entity knowledge like the majority of existing K-PLMs, our approach captures relational knowledge, which contributes to better-improving sentence-level text classification and text matching tasks that play a key role in question answering (QA). We conducted a set of experiments on five text classification tasks and three text matching tasks from three domains, namely E-commerce, Government, and Film&TV, and performed online A/B tests in E-commerce. Experimental results show that our approach is able to achieve substantial improvement on sentence-level question answering tasks and bring beneficial business value in industrial settings.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Fu Sun (3 papers)
  2. Feng-Lin Li (16 papers)
  3. Ruize Wang (11 papers)
  4. Qianglong Chen (25 papers)
  5. Xingyi Cheng (20 papers)
  6. Ji Zhang (176 papers)
Citations (4)