Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
72 tokens/sec
GPT-4o
61 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Explainable Automated Fact-Checking for Public Health Claims (2010.09926v1)

Published 19 Oct 2020 in cs.CL and cs.AI

Abstract: Fact-checking is the task of verifying the veracity of claims by assessing their assertions against credible evidence. The vast majority of fact-checking studies focus exclusively on political claims. Very little research explores fact-checking for other topics, specifically subject matters for which expertise is required. We present the first study of explainable fact-checking for claims which require specific expertise. For our case study we choose the setting of public health. To support this case study we construct a new dataset PUBHEALTH of 11.8K claims accompanied by journalist crafted, gold standard explanations (i.e., judgments) to support the fact-check labels for claims. We explore two tasks: veracity prediction and explanation generation. We also define and evaluate, with humans and computationally, three coherence properties of explanation quality. Our results indicate that, by training on in-domain data, gains can be made in explainable, automated fact-checking for claims which require specific expertise.

Explainable Automated Fact-Checking for Public Health Claims

The paper "Explainable Automated Fact-Checking for Public Health Claims" by Kotonya and Toni addresses the limitations in the current landscape of automated fact-checking models, particularly focusing on two prominent issues: the narrow emphasis on political claims and the deficiency of explainable systems. The research posits the necessity for expertise-driven fact-checking models capable of addressing claims in complex domains like public health. To substantiate their research, the authors introduce the PubHealth dataset, comprising 11.8K public health-related claims paired with journalist-crafted explanations that delineate the rationale behind each claim's veracity label.

Objectives and Contributions

The paper delineates two primary research tasks: veracity prediction and explanation generation. The authors aim to develop a system capable of not only predicting the veracity of claims but also providing understandable explanations, particularly for audiences lacking domain-specific knowledge. The paper's key contributions include:

  1. Dataset Creation: The introduction of PubHealth, a novel dataset tailored for the evaluation of fact-checking systems in the public health domain, distinguished by its incorporation of gold-standard journalistic explanations.
  2. Methodological Framework: The authors present a framework leveraging domain-specific training data to enhance the performance of fact-checking models in veracity prediction and explanation generation.
  3. Coherence Properties for Explanation Evaluation: The paper proposes three coherence properties—strong global coherence, weak global coherence, and local coherence—crafted for evaluating the fidelity of explanations provided by the system.

Methodological Innovations

The approach utilizes advanced NLP techniques for both the prediction and explanation generation tasks. For veracity prediction, models like SciBERT and BioBERT are fine-tuned using in-domain data to enhance accuracy. Moreover, Sentence-BERT (S-BERT) is employed for the task of evidence selection prior to veracity prediction. Meanwhile, explanation generation is modeled using a joint extractive-abstractive summarization method, to ensure explanations are both contextually relevant and comprehensible to non-expert audiences.

Evaluation and Results

The paper employs comprehensive evaluation strategies, encompassing both human and automated approaches. The veracity prediction models, particularly SciBERT and BioBERT v1.1, demonstrated superior performance with significant improvements in F1, precision, and accuracy measures over non-domain-specific models. In explanation generation, the ExplainerFC-Expert model trained with domain-specific data outperformed general models based on ROUGE metrics. Crucially, the paper underscores the effectiveness of using Natural Language Inference (NLI) models to approximate human evaluation of explanation coherence, with a particular utility observed in weak global coherence and local coherence properties.

Implications and Future Directions

The insights from this paper underscore the potential of domain-specific models trained on tailored datasets to significantly enhance both the predictive power and trustworthiness of automated fact-checking systems. The research presents a compelling case for the broader adoption of explainable AI systems, particularly in domains where the implications of misinformation can have severe real-world consequences, such as public health.

Looking forward, the authors suggest expanding this research agenda to encompass additional specialized domains. Moreover, they advocate for a deeper exploration of the alignment between veracity predictions and the quality of generated explanations, which could further bolster the credibility and utility of automated fact-checking systems.

Understanding and improving the coherence and comprehensibility of AI-generated explanations continue to be pivotal challenges in fostering trust and efficacy in AI systems, which this research systematically begins to address.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Neema Kotonya (9 papers)
  2. Francesca Toni (96 papers)
Citations (221)