Can LLMs detect originality, rigour, and impact across fields?
Evaluate the capability, accuracy, and biases of large language models to detect and assess originality, methodological rigour, and impact in journal articles across diverse disciplines, relative to expert judgements.
References
It is not clear whether LLMs could be reasonably effective at detecting originality, rigour, or impact in any or all fields.
— Quantitative Methods in Research Evaluation Citation Indicators, Altmetrics, and Artificial Intelligence
(2407.00135 - Thelwall, 28 Jun 2024) in Section 13.4