Understanding Fine-grained Distortions in Reports of Scientific Findings
Abstract: Distorted science communication harms individuals and society as it can lead to unhealthy behavior change and decrease trust in scientific institutions. Given the rapidly increasing volume of science communication in recent years, a fine-grained understanding of how findings from scientific publications are reported to the general public, and methods to detect distortions from the original work automatically, are crucial. Prior work focused on individual aspects of distortions or worked with unpaired data. In this work, we make three foundational contributions towards addressing this problem: (1) annotating 1,600 instances of scientific findings from academic papers paired with corresponding findings as reported in news articles and tweets wrt. four characteristics: causality, certainty, generality and sensationalism; (2) establishing baselines for automatically detecting these characteristics; and (3) analyzing the prevalence of changes in these characteristics in both human-annotated and large-scale unlabeled data. Our results show that scientific findings frequently undergo subtle distortions when reported. Tweets distort findings more often than science news reports. Detecting fine-grained distortions automatically poses a challenging task. In our experiments, fine-tuned task-specific models consistently outperform few-shot LLM prompting.
- Writing strategies for science communication: Data and computational analysis. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 5327–5344, Online. Association for Computational Linguistics.
- Christopher Bagdon. 2023. You are and expert annotator: Automated annotations with generative models and BWS.
- Scibert: A pretrained language model for scientific text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, pages 3613–3618. Association for Computational Linguistics.
- Beyond facts–a survey and conceptualisation of claims in online discourse analysis. Semantic Web – Interoperability, Usability, Applicability, 13(5):793–827.
- The association between exaggeration in health-related science news and academic press releases: a replication study. Wellcome open research, 4.
- Kathi Canese and Sarah Weis. 2013. PubMed: the Bibliographic Database. The NCBI handbook, 2(1).
- Scientific Research in News Media: A Case Study of Misrepresentation, Sensationalism and Harmful Recommendations. Journal of Science Communication, 21(1):A06.
- Baruch Fischhoff. 2012. Communicating uncertainty fulfilling the duty to inform. Issues in Science and Technology, 28(4):63–70.
- A survey on automated fact-checking. Transactions of the Association for Computational Linguistics, 10:178–206.
- P Sol Hart and Lauren Feldman. 2016. The impact of climate change–related imagery and text on public opinion and behavior change. Science Communication, 38(4):415–441.
- Learning whom to trust with MACE. In Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 1120–1130, Atlanta, Georgia. Association for Computational Linguistics.
- George Hripcsak and Adam S. Rothschild. 2005. Agreement, the f-measure, and reliability in information retrieval. Journal of the American Medical Informatics Association, 12(3):296–298. _eprint: https://academic.oup.com/jamia/article-pdf/12/3/296/2429751/12-3-296.pdf.
- Mistral 7b.
- Mixtral of experts.
- Svetlana Kiritchenko and Saif Mohammad. 2017. Best-worst scaling more reliable than rating scales: A case study on sentiment intensity annotation. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 465–470, Vancouver, Canada. Association for Computational Linguistics.
- Lauren M Kuehne and Julian D Olden. 2015. Lay summaries needed to enhance science communication. Proceedings of the National Academy of Sciences, 112(12):3585–3586.
- The effects of scientific messages and narratives about vaccination. PLoS One, 16(3):e0248328.
- Revise and Resubmit: An Intertextual Model of Text-based Collaboration in Peer Review. Comput. Linguistics, 48(4):949–986.
- John Lawrence and Chris Reed. 2019. Argument mining: A survey. Computational Linguistics, 45(4):765–818.
- Roberta: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692.
- S2ORC: the semantic scholar open research corpus. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020, pages 4969–4983. Association for Computational Linguistics.
- Mary L McHugh. 2012. Interrater reliability: the kappa statistic. Biochem Med (Zagreb), 22(3):276–282.
- Covert: A corpus of fact-checked biomedical COVID-19 tweets. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, LREC 2022, Marseille, France, 20-25 June 2022, pages 244–257. European Language Resources Association.
- Potato: The portable text annotation tool. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations.
- Jiaxin Pei and David Jurgens. 2021a. Measuring sentence-level and aspect-level (un)certainty in science communications. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 9959–10011, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Jiaxin Pei and David Jurgens. 2021b. Measuring sentence-level and aspect-level (un)certainty in science communications. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 9959–10011, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- David F Ransohoff and Richard M Ransohoff. 2001. Sensationalism in the Media: When Scientists and Journalists May be Complicit Collaborators. Effective clinical practice, 4(4).
- Joselita T Salita. 2015. Writing for lay audiences: A challenge for scientists. Medical Writing, 24(4):183–189.
- The association between exaggeration in health related science news and academic press releases: retrospective observational study. Bmj, 349.
- Mass Communication Systems and Communication Accuracy in Science News Reporting. Journalism Quarterly, 47(4):673–683.
- Llama 2: Open foundation and fine-tuned chat models.
- Infolossqa: Characterizing and recovering information loss in text simplification.
- Juraj Vladika and Florian Matthes. 2023. Scientific fact-checking: A survey of resources and approaches. In Findings of the Association for Computational Linguistics: ACL 2023, pages 6215–6230, Toronto, Canada. Association for Computational Linguistics.
- Fact or fiction: Verifying scientific claims. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP 2020, Online, November 16-20, 2020, pages 7534–7550. Association for Computational Linguistics.
- Dustin Wright and Isabelle Augenstein. 2021a. CiteWorth: Cite-Worthiness Detection for Improved Scientific Document Understanding. In Findings of the Association for Computational Linguistics: ACL/IJCNLP 2021, Online Event, August 1-6, 2021, volume ACL/IJCNLP 2021 of Findings of ACL, pages 1796–1807. Association for Computational Linguistics.
- Dustin Wright and Isabelle Augenstein. 2021b. Semi-supervised exaggeration detection of health science press releases. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 10824–10836, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Modeling information change in science communication with semantically matched paraphrases. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 1783–1807, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Detecting causal language use in science findings. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 4664–4674, Hong Kong, China. Association for Computational Linguistics.
- Measuring correlation-to-causation exaggeration in press releases. In Proceedings of the 28th International Conference on Computational Linguistics, pages 4860–4872, Barcelona, Spain (Online). International Committee on Computational Linguistics.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.