Leveraging LLMs for Scalable Misinformation Correction on Social Media
Introduction
The proliferation of misinformation on social media platforms poses a significant challenge to public perception, often undermining trust in science and democracy. Traditional methods of misinformation correction involve expert and layperson intervention, which although effective, cannot scale to address the volume of misinformation generated daily. This limitation is exacerbated by the evolution of LLMs that, while facilitating misinformation creation, also bear potential for scalable misinformation correction. The paper introduces Muse, a novel approach utilizing an LLM augmented with access to and credibility evaluation of up-to-date information for multimodal misinformation correction on social media. Muse demonstrates superior performance in correcting misinformation compared to GPT-4 and high-quality corrections from laypeople.
Approach
Muse's design enables it to address not only textual but also visual misinformation through the integration of visuals, up-to-date factual, and credible web knowledge retrieval. The process begins with image descriptions utilizing image captioning models enhanced by celebrity recognition and Optical Character Recognition (OCR) for more informative interpretation of visual content. For textual misinformation, relevant web pages are retrieved using generated queries and filtered based on direct relevance and credibility evaluation. Muse generates corrections leveraging extracted evidence from web pages, ensuring accurate, trustworthy references and explanations.
Evaluation
An extensive evaluation involving experts in fact-checking and journalism assessed corrections generated by Muse across 13 dimensions. These included the factuality of explanation, relevance and credibility of references, and overall quality of corrections among others. Muse's corrections outperformed those generated by GPT-4 by 37% and laypeople's high-quality corrections by 29%, showcasing its capability to promptly correct misinformation after it appears on social media. Particularly, Muse excelled in identifying inaccuracies, generating relevant and factual text, and providing credible references.
Implications and Future Directions
The findings underscore the potential of Muse and similar technologies in combating misinformation on social media platforms both effectively and efficiently. The approach outlined not only addresses the scalability issue faced by manual corrections but also introduces a method to enhance the accuracy and trustworthiness of generated corrections. Future research could explore the integration of video inputs, application across multiple languages, and extend the evaluation to other platforms beyond X Community Notes. Moreover, further developments might aim to reduce the correction generation cost and time, already estimated at $0.5 per social media post, and examine the impact of correction immediacy on Muse's performance.
Conclusion
The development of Muse represents a significant advancement in the use of LLMs for misinformation correction on social media. By integrating capabilities for handling multimodal misinformation, accessing up-to-date information, and generating corrections with accurate references, Muse sets a new standard for automated misinformation correction technologies. Its superior performance, demonstrated through a comprehensive expert evaluation, highlights its potential as a scalable solution to the misinformation problem that plagues social media platforms.