Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MultiFC: A Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims (1909.03242v2)

Published 7 Sep 2019 in cs.CL, cs.IR, cs.LG, and stat.ML

Abstract: We contribute the largest publicly available dataset of naturally occurring factual claims for the purpose of automatic claim verification. It is collected from 26 fact checking websites in English, paired with textual sources and rich metadata, and labelled for veracity by human expert journalists. We present an in-depth analysis of the dataset, highlighting characteristics and challenges. Further, we present results for automatic veracity prediction, both with established baselines and with a novel method for joint ranking of evidence pages and predicting veracity that outperforms all baselines. Significant performance increases are achieved by encoding evidence, and by modelling metadata. Our best-performing model achieves a Macro F1 of 49.2%, showing that this is a challenging testbed for claim veracity prediction.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Isabelle Augenstein (131 papers)
  2. Christina Lioma (66 papers)
  3. Dongsheng Wang (47 papers)
  4. Lucas Chaves Lima (8 papers)
  5. Casper Hansen (22 papers)
  6. Christian Hansen (51 papers)
  7. Jakob Grue Simonsen (43 papers)
Citations (230)

Summary

  • The paper presents MultiFC, the largest natural dataset with 34,918 claims from 26 fact-checking websites for evidence-based verification.
  • It employs a Multi-Task Learning framework to reconcile diverse labeling schemes, achieving a Macro F1 score of 49.2% in claim veracity prediction.
  • The findings enable enhanced automated fact-checking tools and provide a robust benchmark for advancing research in NLP and misinformation detection.

MultiFC: A Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims

The paper, "MultiFC: A Real-World Multi-Domain Dataset for Evidence-Based Fact Checking of Claims," introduces a substantial dataset aimed at advancing the field of automatic claim verification. The dataset, MultiFC, is the largest publicly available set of naturally occurring factual claims. It consists of 34,918 claims collected from 26 different fact-checking websites in English, paired with associated textual sources and metadata, with veracity labels assigned by human expert journalists. This work responds to the growing need for combating misinformation and disinformation, particularly prevalent in social media ecosystems.

Dataset Composition and Characteristics

The MultiFC dataset distinguishes itself by harnessing real-world data sources compared to other smaller and sometimes artificially constructed datasets. It incorporates:

  • Claims collected from diverse fact-checking websites like politifact.com and snopes.com, among others.
  • Rich metadata including claim source, claim speaker, publication date, and related URLs.
  • Labeling of claim veracity handled by journalist experts, dividing claims into a comprehensive set of labels (between 2 and 40) depending on the source domain.

A notable challenge addressed in the paper is the varied labeling schemes across different fact-checking portals, which the authors tackle by employing a Multi-Task Learning (MTL) framework. This approach enables modeling disparate label spaces across domains effectively.

Technical Contributions and Results

The core of the technical innovation lies in the proposed frameworks for automatic veracity prediction. The authors evaluated several machine learning models, employing both existing baselines and a novel method involving the joint ranking of evidence pages followed by veracity prediction. They identify a significant improvement when incorporating the semantic encoding of evidence and modeling metadata.

The best-performing model achieves a Macro F1 score of 49.2%, which underscores the challenging nature of claim veracity prediction using this dataset. This performance metric implies meaningful strides towards refining automatic claim verification processes.

Implications and Future Directions

This research bears several pertinent implications:

  • Practical Applications: The methodologies and findings can be leveraged to enhance automated fact-checking tools, thereby aiding journalists and social media platforms in identifying and curbing misinformation quickly and efficiently.
  • Research Advancements: The dataset serves as a robust benchmark for the development and testing of more sophisticated claim verification algorithms, encouraging advancements in fields like NLP and machine learning.

The authors suggest several future avenues, such as improving evidence page encoding and integrating more comprehensive temporal data, which could further augment the veracity prediction tasks. Additionally, exploring joint learning setups for related tasks, like stance detection, could yield more insightful relationships between claims and their supporting or opposing evidence.

In Summary

The MultiFC dataset and the analytical methodologies proposed by the authors mark a significant contribution to the literature on automatic claim verification. They offer a standardized testbed for developing and benchmarking future models in this domain while addressing real-world complexities posed by diverse labeling standards and varied data sources. The research opens pathways for improved fact-checking tools that can scale effectively with the ever-growing content on social platforms and news media.