Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RAGAR, Your Falsehood Radar: RAG-Augmented Reasoning for Political Fact-Checking using Multimodal Large Language Models (2404.12065v2)

Published 18 Apr 2024 in cs.CL, cs.AI, cs.CY, cs.ET, and cs.MA

Abstract: The escalating challenge of misinformation, particularly in political discourse, requires advanced fact-checking solutions; this is even clearer in the more complex scenario of multimodal claims. We tackle this issue using a multimodal LLM in conjunction with retrieval-augmented generation (RAG), and introduce two novel reasoning techniques: Chain of RAG (CoRAG) and Tree of RAG (ToRAG). They fact-check multimodal claims by extracting both textual and image content, retrieving external information, and reasoning subsequent questions to be answered based on prior evidence. We achieve a weighted F1-score of 0.85, surpassing a baseline reasoning technique by 0.14 points. Human evaluation confirms that the vast majority of our generated fact-check explanations contain all information from gold standard data.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. A survey on multimodal disinformation detection. In Proceedings of the 29th International Conference on Computational Linguistics, pages 6625–6643, Gyeongju, Republic of Korea. International Committee on Computational Linguistics.
  2. Self-RAG: Learning to retrieve, generate, and critique through self-reflection. In The Twelfth International Conference on Learning Representations.
  3. Fake news, disinformation and misinformation in social media: a review. Social Network Analysis and Mining, 13(1):30.
  4. FacTool: Factuality detection in generative AI – a tool augmented framework for multi-task and multi-domain scenarios. arXiv preprint arXiv:2307.13528.
  5. The state of human-centered NLP technology for fact-checking. Information Processing & Management, 60(2):103219.
  6. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
  7. Chain-of-verification reduces hallucination in large language models. arXiv preprint arXiv:2309.11495.
  8. Texts as images in prompt tuning for multi-label image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2808–2817.
  9. A richly annotated corpus for different tasks in automated fact-checking. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 493–503, Hong Kong, China. Association for Computational Linguistics.
  10. Andrew F. Hayes and Klaus Krippendorff. 2007. Answering the call for a standard reliability measure for coding data. Communication Methods and Measures, 1:77 – 89.
  11. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, Online. Association for Computational Linguistics.
  12. OpenAI. 2023. Gpt-4v: A multimodal transformer for vision and language.
  13. Fact-checking complex claims with program-guided reasoning. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 6981–7004, Toronto, Canada. Association for Computational Linguistics.
  14. Check your facts and try again: Improving large language models with external knowledge and automated feedback. arXiv preprint arXiv:2302.12813.
  15. Learning transferable visual models from natural language supervision. In Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 8748–8763. PMLR.
  16. The spread of true and false news online. Science, 359(6380):1146–1151.
  17. End-to-end multimodal fact-checking and explanation generation: A challenging dataset and models. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM.
  18. End-to-end multimodal fact-checking and explanation generation: A challenging dataset and models. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’23, page 2733–2743, New York, NY, USA. Association for Computing Machinery.
  19. Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601.
  20. On the origins of memes by means of fringe web communities. In Proceedings of the Internet Measurement Conference 2018, IMC ’18, page 188–202, New York, NY, USA. Association for Computing Machinery.
  21. Fengzhu Zeng and Wei Gao. 2024. Justilm: Few-shot justification generation for explainable fact-checking of real-world claims. arXiv preprint arXiv:2401.08026.
  22. Xuan Zhang and Wei Gao. 2023. Towards llm-based fact verification on news claims with a hierarchical step-by-step prompting method. arXiv preprint arXiv:2310.00305.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. M. Abdul Khaliq (1 paper)
  2. P. Chang (247 papers)
  3. M. Ma (18 papers)
  4. B. Pflugfelder (1 paper)
  5. F. Miletić (1 paper)
Citations (2)

Summary

We haven't generated a summary for this paper yet.