Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 46 tok/s Pro
GPT-5 Medium 23 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 101 tok/s Pro
Kimi K2 179 tok/s Pro
GPT OSS 120B 435 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Language-Agnostic Modeling of Source Reliability on Wikipedia (2410.18803v2)

Published 24 Oct 2024 in cs.SI and cs.LG

Abstract: Over the last few years, content verification through reliable sources has become a fundamental need to combat disinformation. Here, we present a language-agnostic model designed to assess the reliability of sources across multiple language editions of Wikipedia. Utilizing editorial activity data, the model evaluates source reliability within different articles of varying controversiality such as Climate Change, COVID-19, History, Media, and Biology topics. Crafting features that express domain usage across articles, the model effectively predicts source reliability, achieving an F1 Macro score of approximately 0.80 for English and other high-resource languages. For mid-resource languages, we achieve 0.65 while the performance of low-resource languages varies; in all cases, the time the domain remains present in the articles (which we dub as permanence) is one of the most predictive features. We highlight the challenge of maintaining consistent model performance across languages of varying resource levels and demonstrate that adapting models from higher-resource languages can improve performance. This work contributes not only to Wikipedia's efforts in ensuring content verifiability but in ensuring reliability across diverse user-generated content in various language communities.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Pablo Aragón and Diego Sáez-Trumper. 2021. A preliminary approach to knowledge integrity risk assessment in Wikipedia projects. CoRR abs/2106.15940 (2021), 1–4. arXiv:2106.15940 https://arxiv.org/abs/2106.15940
  2. The Dynamics of (Not) Unfollowing Misinformation Spreaders. In Proceedings of the ACM on Web Conference 2024. 1115–1125.
  3. Longitudinal Assessment of Reference Quality on Wikipedia. In Proceedings of the ACM Web Conference 2023 (WWW ’23). ACM, New York, NY, USA, 2831–2839. https://doi.org/10.1145/3543507.3583218
  4. A Comparative Study of Reference Reliability in Multiple Language Editions of Wikipedia. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM ’23). ACM, New York, NY, USA, 3743–3747. https://doi.org/10.1145/3583780.3615254
  5. Golding Barret. 2021. Iffy index of unreliable sources. (2021). https://iffy.news/index/ [accessed 2024 April 3].
  6. Societal Controversies in Wikipedia Articles. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI ’15). Association for Computing Machinery, New York, NY, USA, 193–196. https://doi.org/10.1145/2702123.2702436
  7. Citation detective: a public dataset to improve and quantify wikipedia citation quality at scale. Wiki Workshop.
  8. Joint Estimation of User And Publisher Credibility for Fake News Detection. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management (CIKM ’20). Association for Computing Machinery, New York, NY, USA, 1993–1996. https://doi.org/10.1145/3340531.3412066
  9. Noam Cohen. 2021. One Woman’s Mission to Rewrite Nazi History on Wikipedia. WIRED. [Online; accessed 15-Apr-2024].
  10. Language-Agnostic Modeling of Wikipedia Articles for Content Quality Assessment across Languages. Proceedings of the International AAAI Conference on Web and Social Media 18, 01 (2024), 1–11.
  11. The spreading of misinformation online. Proceedings of the national academy of Sciences 113, 3 (2016), 554–559.
  12. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).
  13. Introducing an “invisible enemy”: a case study of knowledge construction regarding microplastics in Japanese Wikipedia. New Media & Society 0, 0 (2023), 14614448221149747.
  14. Multilingual entity linking system for Wikipedia with a machine-in-the-loop approach. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management. ACM, New York, NY, USA, 3818–3827.
  15. Interpretable Fake News Detection with Graph Evidence. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM ’23). Association for Computing Machinery, New York, NY, USA, 659–668. https://doi.org/10.1145/3583780.3614936
  16. Edit-History Vis: An Interactive Visual Exploration and Analysis on Wikipedia Edit History. In 2023 IEEE 16th Pacific Visualization Symposium (PacificVis). IEEE, New York, NY, USA, 157–166.
  17. Scott A Hale. 2014. Multilinguals and Wikipedia editing. In Proceedings of the 2014 ACM conference on Web science. 99–108.
  18. Aaron Halfaker and R Stuart Geiger. 2020. ORES: Lowering barriers with participatory machine learning in wikipedia. Proceedings of the ACM on Human-Computer Interaction 4, CSCW2 (2020), 1–37.
  19. Language-agnostic Topic Classification for Wikipedia. In Companion Proceedings of the Web Conference 2021 (WWW ’21). Association for Computing Machinery, New York, NY, USA, 594–601. https://doi.org/10.1145/3442442.3452347
  20. Governance Capture in a Self-Governing Community: A Qualitative Comparison of the Serbo-Croatian Wikipedias. arXiv preprint arXiv:2311.03616 (2023).
  21. From causes to consequences, from chat to crisis. The different climate changes of science and Wikipedia. Environmental Science & Policy 148 (2023), 103553. https://doi.org/10.1016/j.envsci.2023.103553
  22. Disinformation on the Web: Impact, Characteristics, and Detection of Wikipedia Hoaxes. In Proceedings of the 25th International Conference on World Wide Web (WWW ’16). International World Wide Web Conferences Steering Committee, Republic and Canton of Geneva, CHE, 591–602. https://doi.org/10.1145/2872427.2883085
  23. Templates and Trust-o-meters: Towards a widely deployable indicator of trust in Wikipedia. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 1–17.
  24. Multilingual ranking of Wikipedia articles with quality and popularity assessment in different topics. Computers 8, 3 (2019), 60.
  25. Modeling popularity and reliability of sources in multilingual Wikipedia. Information 11, 5 (2020), 263.
  26. Scott M Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Advances in Neural Information Processing Systems 30, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.). Curran Associates, Inc., Red Hook, NY, USA, 4765–4774. http://papers.nips.cc/paper/7062-a-unified-approach-to-interpreting-model-predictions.pdf
  27. Improving wikipedia verifiability with ai. Nature Machine Intelligence 5, 10 (2023), 1142–1148.
  28. Citation Needed: A Taxonomy and Algorithmic Assessment of Wikipedia’s Verifiability. In The World Wide Web Conference (WWW ’19). Association for Computing Machinery, New York, NY, USA, 1567–1578. https://doi.org/10.1145/3308558.3313618
  29. Wikipedia: a self-organizing bureaucracy. Information, Communication & Society 26, 7 (2023), 1285–1302.
  30. Aaron Shaw and Benjamin M Hill. 2014. Laboratories of oligarchy? How the iron law extends to peer production. Journal of Communication 64, 2 (2014), 215–238.
  31. Studying Fake News via Network Analysis: Detection and Mitigation. In Emerging Research Challenges and Opportunities in Computational Social Network Analysis and Mining, Nitin Agarwal, Nima Dokoohaki, and Serpil Tokdemir (Eds.). Springer International Publishing, Cham, 43–65. https://doi.org/10.1007/978-3-319-94105-9_3
  32. Mining user-aware multi-relations for fake news detection in large scale online social networks. In Proceedings of the sixteenth ACM international conference on web search and data mining. 51–59.
  33. A commonsense-infused language-agnostic learning framework for enhancing prediction of political bias in multilingual news headlines. Knowledge-Based Systems 277 (2023), 110838.
  34. Fair Multilingual Vandalism Detection System for Wikipedia. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’23). Association for Computing Machinery, New York, NY, USA, 4981–4990. https://doi.org/10.1145/3580305.3599823
  35. Mykola Trokhymovych and Diego Saez-Trumper. 2021. WikiCheck: An End-to-end Open Source Automatic Fact-Checking API based on Wikipedia. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management (CIKM ’21). Association for Computing Machinery, New York, NY, USA, 4155–4164. https://doi.org/10.1145/3459637.3481961
  36. The spread of true and false news online. science 359, 6380 (2018), 1146–1151.
  37. Krzysztof W\kecel and Włodzimierz Lewoniewski. 2015. Modelling the quality of attributes in Wikipedia infoboxes. In Business Information Systems Workshops: BIS 2015 International Workshops, Poznań, Poland, June 24-26, 2015, Revised Papers 18. Springer International Publishing, Cham, 308–320.
  38. Wikipedia contributors. 2024. English Wikipedia. 2024. Wikipedia:Reliable sources/Perennial sources. https://en.wikipedia.org/wiki/Wikipedia:Reliable_sources/Perennial_sources [Online; accessed 12-April-2024].
  39. The FAIR Guiding Principles for scientific data management and stewardship. Scientific data 3, 1 (2016), 1–9.
  40. Prompt-and-Align: Prompt-Based Social Alignment for Few-Shot Fake News Detection. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM ’23). Association for Computing Machinery, New York, NY, USA, 2726–2736. https://doi.org/10.1145/3583780.3615015
  41. MSynFD: Multi-hop Syntax aware Fake News Detection. In Proceedings of the ACM on Web Conference 2024. 4128–4137.
  42. HiPo: Detecting Fake News via Historical and Multi-Modal Analyses of Social Media Posts. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management (CIKM ’23). Association for Computing Machinery, New York, NY, USA, 2805–2815. https://doi.org/10.1145/3583780.3614914
  43. Identifying cost-effective debunkers for multi-stage fake news mitigation campaigns. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining. 1206–1214.
  44. Puyu Yang and Giovanni Colavizza. 2024. Polarization and reliability of news sources in Wikipedia. Online Information Review ahead-of-print, ahead-of-print (2024), 1–18.
  45. Junting Ye and Steven Skiena. 2019. MediaRank: Computational Ranking of Online News Sources. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ’19). Association for Computing Machinery, New York, NY, USA, 2469–2477. https://doi.org/10.1145/3292500.3330709
  46. Dave Van Zandt. 2024. Media Bias/Fact-Check. (2024). https://mediabiasfactcheck.com/ [accessed 2024 April 23].
  47. M3Exam: A Multilingual, Multimodal, Multilevel Benchmark for Examining Large Language Models. In Advances in Neural Information Processing Systems, A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine (Eds.), Vol. 36. Curran Associates, Inc., New York, 5484–5505. https://proceedings.neurips.cc/paper_files/paper/2023/file/117c5c8622b0d539f74f6d1fb082a2e9-Paper-Datasets_and_Benchmarks.pdf
  48. Don’t trust ChatGPT when your question is not in English: A study of multilingual abilities and types of LLMs. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Singapore, 7915–7927.
  49. Detecting health misinformation in online health communities: Incorporating behavioral features into machine learning based approaches. Information Processing & Management 58, 1 (2021), 102390.
  50. Gender and country biases in Wikipedia citations to scholarly publications. Journal of the Association for Information Science and Technology 74, 2 (2023), 219–233.

Summary

  • The paper introduces a model that uses editorial activity data to predict Wikipedia source reliability without relying on language-specific processing.
  • It demonstrates strong performance metrics, achieving an F1 Macro score of approximately 0.80 for high-resource languages while noting lower scores in mid- and low-resource contexts.
  • The findings support enhanced editorial guidelines and offer a pathway for adapting models to improve verifiability and combat misinformation on multilingual platforms.

Language-Agnostic Modeling of Source Reliability on Wikipedia

The paper "Language-Agnostic Modeling of Source Reliability on Wikipedia" presents an innovative approach towards evaluating the reliability of sources cited on Wikipedia, using a model that transcends language barriers. Given the increasing significance of ensuring content verifiability in combatting disinformation, this paper addresses a critical challenge: assessing the reliability of sources across diverse language editions of Wikipedia.

Summary of Methodology and Key Findings

The authors introduce a model leveraging editorial activity data—a unique idea that circumvents the need for language-specific processing—to estimate the reliability of sources cited in various Wikipedia articles, encompassing topics like Climate Change, COVID-19, Biology, History, and Media. Their approach focuses on crafting language-agnostic features, such as the permanence of sources within articles, the number of articles referencing a source, and the number of unique users interacting with the source citations.

Key performance metrics of their model, such as the F1 Macro score, reveal an impressively effective prediction of source reliability in high-resource language settings, with scores reaching approximately 0.80 for English and other high-resource languages. However, performance varies notably in mid-resource (F1 Macro 0.65) and low-resource languages, with the latter witnessing more fluctuation, underscoring the challenge of consistent performance across diverse linguistic contexts.

Strong Numerical Results and Insights

One of the strongest numerical results is the model's capability to allude to source reliability through editorial behavior patterns, achieving significant predictive power with language-agnostic features alone. Notably, permanence, a key feature indicating the duration a source remains cited in articles, emerged as among the most pivotal in predicting reliability.

Furthermore, the model demonstrates adaptability potential; while its performance decreases when applied to cross-language and cross-topic settings, improvements are observed when models trained in high-resource languages are adapted to mid- and low-resource languages. This highlights the model's versatility and potential to be refined for broader applicability.

Implications and Future Work

The paper contributes significantly to the efforts in maintaining Wikipedia's role as a reliable information source by providing a mechanism that can enhance editorial monitoring algorithms, primarily in language editions lacking substantial resources. The authors demonstrate that using a mix of language-derived data can improve models for low-resource languages, suggesting future research should continue to explore cross-language and cross-topic machine learning model adaptations.

A notable implication of the findings is the possibility of using the model to support editorial guidelines and help curate community-maintained reliable and unreliable source lists, crucial for Wikipedia's information integrity. Future advancements could involve augmenting this language-agnostic model with semantic linking or user session data to improve its predictive accuracy and applicability.

Conclusion

In introducing a model that astute editors can use across Wikipedia's wide array of languages, the paper tackles the pressing challenge of source reliability in an increasingly multilingual and multifacetal online ecosystem. The authors successfully provide a foundational approach to discerning reliable sources using a language-neutral framework, setting the stage for further developments in AI models aimed at content reliability and verifiability. These advancements not only bolster the integrity of Wikipedia but have wider implications for combating misinformation across user-generated content platforms.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 22 likes.

Upgrade to Pro to view all of the tweets about this paper: