Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

PRISMA-DFLLM: An Extension of PRISMA for Systematic Literature Reviews using Domain-specific Finetuned Large Language Models (2306.14905v1)

Published 15 Jun 2023 in cs.CL and cs.AI

Abstract: With the proliferation of open-sourced LLMs and efficient finetuning techniques, we are on the cusp of the emergence of numerous domain-specific LLMs that have been finetuned for expertise across specialized fields and applications for which the current general-purpose LLMs are unsuitable. In academia, this technology has the potential to revolutionize the way we conduct systematic literature reviews (SLRs), access knowledge and generate new insights. This paper proposes an AI-enabled methodological framework that combines the power of LLMs with the rigorous reporting guidelines of the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA). By finetuning LLMs on domain-specific academic papers that have been selected as a result of a rigorous SLR process, the proposed PRISMA-DFLLM (for Domain-specific Finetuned LLMs) reporting guidelines offer the potential to achieve greater efficiency, reusability and scalability, while also opening the potential for conducting incremental living systematic reviews with the aid of LLMs. Additionally, the proposed approach for leveraging LLMs for SLRs enables the dissemination of finetuned models, empowering researchers to accelerate advancements and democratize cutting-edge research. This paper presents the case for the feasibility of finetuned LLMs to support rigorous SLRs and the technical requirements for realizing this. This work then proposes the extended PRISMA-DFLLM checklist of reporting guidelines as well as the advantages, challenges, and potential implications of implementing PRISMA-DFLLM. Finally, a future research roadmap to develop this line of AI-enabled SLRs is presented, paving the way for a new era of evidence synthesis and knowledge discovery.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. OpenAI: About. https://openai.com/about/. Accessed on 14 June 2023.
  2. Stanford crfm introduces pubmedgpt 2.7b. https://hai.stanford.edu/news/stanford-crfm-introduces-pubmedgpt-27b, 2022. Accessed: 13 June 2023.
  3. Language models are few-shot learners. arXiv preprint arXiv:2005.14165, 2020.
  4. Parameter-efficient fine-tuning design spaces. arXiv preprint arXiv:2301.01821, 2023.
  5. D. Curcic. Number of academic papers published per year. https://wordsrated.com/number-of-academic-papers-published-per-year. Accessed on 14th June 2023.
  6. Qlora: Efficient finetuning of quantized llms, 2023.
  7. Living systematic reviews: an emerging opportunity to narrow the evidence-practice gap. PLoS Med, 11(2):e1001603, 2014.
  8. The use of text-mining software to facilitate screening of literature on centredness in health care. Systematic Reviews, 12(1):73, 2023.
  9. G-adapter: Towards structure-aware parameter-efficient transfer learning for graph transformer networks. arXiv preprint arXiv:2305.10329, 2023.
  10. Lora: Low-rank adaptation of large language models, 2021.
  11. Llm-adapters: An adapter family for parameter-efficient fine-tuning of large language models. arXiv preprint arXiv:2304.01933, 2023.
  12. The prisma extension statement for reporting of systematic reviews incorporating network meta-analyses of health care interventions: checklist and explanations. Annals of internal medicine, 162(11):777–784, 2015.
  13. Ensemble of deep learning language models to support the creation of living systematic reviews for the covid-19 literature. Systematic Reviews, 12(1):94, 2023.
  14. Guidance to best tools and practices for systematic reviews. Systematic Reviews, 12(1):96, 2023. ISSN 2046-4053. doi: 10.1186/s13643-023-02255-9. URL https://doi.org/10.1186/s13643-023-02255-9.
  15. E. Landhuis. Scientific literature: Information overload. Nature, 535(7612):457–458, 2016.
  16. Do we still need clinical language models? arXiv preprint arXiv:2302.08091, 2023.
  17. Toward systematic review automation: a practical guide to using machine learning tools in research synthesis. Systematic reviews, 8:1–10, 2019.
  18. Improving the quality of reports of meta-analyses of randomised controlled trials: the quorom statement. The Lancet, 354(9193):1896–1900, 1999.
  19. Preferred reporting items for systematic reviews and meta-analyses: the prisma statement. PLoS medicine, 6(7):e1000097, 2009.
  20. Preferred reporting items for systematic review and meta-analysis protocols (prisma-p) 2015 statement. Systematic reviews, 4(1):1–9, 2015.
  21. The effect of machine learning tools for evidence synthesis on resource use and time-to-completion: protocol for a retrospective pilot study. Systematic Reviews, 12(1):1–8, 2023.
  22. A question of trust: can we build an evidence base to gain trust in systematic review automation technologies? Systematic reviews, 8(1):1–8, 2019.
  23. The prisma 2020 statement: an updated guideline for reporting systematic reviews. International journal of surgery, 88:105906, 2021.
  24. Are chatgpt and large language models “the answer” to bringing us closer to systematic review automation? Systematic Reviews, 12(1):72, 2023.
  25. How to properly use the prisma statement. Systematic Reviews, 10(1):1–3, 2021.
  26. Amstar 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. bmj, 358, 2017.
  27. Societal biases in language generation: Progress and challenges. arXiv preprint arXiv:2105.04054, 2021.
  28. Large language models encode clinical knowledge. arXiv preprint arXiv:2212.13138, 2022.
  29. Towards expert-level medical question answering with large language models. arXiv preprint arXiv:2305.09617, 2023.
  30. Developing prisma-rr, a reporting guideline for rapid reviews of primary studies (protocol). Equator Network, 2018.
  31. Preferred reporting items for systematic review and meta-analyses of individual participant data: the prisma-ipd statement. Jama, 313(16):1657–1665, 2015.
  32. Meta-analysis of observational studies in epidemiology: a proposal for reporting. Jama, 283(15):2008–2012, 2000.
  33. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  34. Prisma extension for scoping reviews (prisma-scr): checklist and explanation. Annals of internal medicine, 169(7):467–473, 2018.
  35. Systematic review automation technologies. Systematic reviews, 3:1–15, 2014.
  36. Adamix: Mixture-of-adaptations for parameter-efficient model tuning. arXiv preprint arXiv:2210.17451, 2022.
  37. Robis: a new tool to assess risk of bias in systematic reviews was developed. Journal of clinical epidemiology, 69:225–234, 2016.
  38. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:2303.17564, 2023.
  39. Lima: Less is more for alignment. arXiv preprint arXiv:2305.11206, 2023.
  40. Prisma harms checklist: improving harms reporting in systematic reviews. bmj, 352, 2016.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Teo Susnjak (23 papers)
Citations (10)