Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Causality-aware Concept Extraction based on Knowledge-guided Prompting (2305.01876v5)

Published 3 May 2023 in cs.CL and cs.AI

Abstract: Concepts benefit natural language understanding but are far from complete in existing knowledge graphs (KGs). Recently, pre-trained LLMs (PLMs) have been widely used in text-based concept extraction (CE). However, PLMs tend to mine the co-occurrence associations from massive corpus as pre-trained knowledge rather than the real causal effect between tokens. As a result, the pre-trained knowledge confounds PLMs to extract biased concepts based on spurious co-occurrence correlations, inevitably resulting in low precision. In this paper, through the lens of a Structural Causal Model (SCM), we propose equipping the PLM-based extractor with a knowledge-guided prompt as an intervention to alleviate concept bias. The prompt adopts the topic of the given entity from the existing knowledge in KGs to mitigate the spurious co-occurrence correlations between entities and biased concepts. Our extensive experiments on representative multilingual KG datasets justify that our proposed prompt can effectively alleviate concept bias and improve the performance of PLM-based CE models.The code has been released on https://github.com/siyuyuan/KPCE.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Siyu Yuan (46 papers)
  2. Deqing Yang (55 papers)
  3. Jinxi Liu (2 papers)
  4. Shuyu Tian (2 papers)
  5. Jiaqing Liang (62 papers)
  6. Yanghua Xiao (151 papers)
  7. Rui Xie (59 papers)
Citations (9)