AI and Generative AI for Research Discovery and Summarization (2401.06795v2)
Abstract: AI and generative AI tools, including chatbots like ChatGPT that rely on LLMs, have burst onto the scene this year, creating incredible opportunities to increase work productivity and improve our lives. Statisticians and data scientists have begun experiencing the benefits from the availability of these tools in numerous ways, such as the generation of programming code from text prompts to analyze data or fit statistical models. One area that these tools can make a substantial impact is in research discovery and summarization. Standalone tools and plugins to chatbots are being developed that allow researchers to more quickly find relevant literature than pre-2023 search tools. Furthermore, generative AI tools have improved to the point where they can summarize and extract the key points from research articles in succinct language. Finally, chatbots based on highly parameterized LLMs can be used to simulate abductive reasoning, which provides researchers the ability to make connections among related technical topics, which can also be used for research discovery. We review the developments in AI and generative AI for research discovery and summarization, and propose directions where these types of tools are likely to head in the future that may be of interest to statistician and data scientists.
- GPT-4 technical report. ArXiv, abs/2303.08774.
- Artificial hallucinations in ChatGPT: Implications in scientific writing. Cureus, 15(2).
- A multitask, multilingual, multimodal evaluation of ChatGPT on reasoning, hallucination, and interactivity. arXiv preprint arXiv:2302.04023.
- Google scholar’s ranking algorithm: An introductory overview. In Proceedings of the 12th international conference on scientometrics and informetrics (ISSI’09), volume 1, pages 230–241. Rio de Janeiro (Brazil).
- Abductive commonsense reasoning. arXiv preprint arXiv:1908.05739.
- Modern multidimensional scaling: Theory and applications. Springer Science & Business Media.
- Perspective: Large language models in applied mechanics. Journal of Applied Mechanics, 90(10):101008.
- Language models are few-shot learners. ArXiv, abs/2005.14165.
- Tldr: Extreme summarization of scientific documents. ArXiv, abs/2004.15011.
- Chincha, D. (2023). Number of ChatGPT plugins. Accessed on December 25, 2023.
- Davis, E. (2023). Mathematics, word problems, common sense, and artificial intelligence. ArXiv, abs/2301.09723.
- Doherty, S. (2016). Translations: The impact of translation technologies on the process and product of translation. International Journal of Communication, 10:23.
- Gower, J. C. (1966). Some distance properties of latent root and vector methods used in multivariate analysis. Biometrika, 53(3-4):325–338.
- Learning to fake it: Limited responses and fabricated references provided by ChatGPT for medical questions. Mayo Clinic Proceedings: Digital Health, 1(3):226–234.
- How close is ChatGPT to human experts? comparison corpus, evaluation, and detection. ArXiv, abs/2301.07597.
- On using monolingual corpora in neural machine translation. ArXiv, abs/1503.03535.
- Towards reasoning in large language models: A survey. ArXiv, abs/2212.10403.
- Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12):1–38.
- Kung, J. (2023). Elicit. The Journal of the Canadian Health Libraries Association, 44:15 – 18.
- Morris, M. R. (2023). Scientists’ perspectives on the potential for generative AI in their fields. ArXiv, abs/2304.01420.
- Scite: A smart citation index that displays the context of citations and classifies their intent using deep learning. bioRxiv.
- Pareschi, R. (2023). Abductive reasoning with the GPT-4 language model: Case studies from criminal investigation, medical practice, scientific research. Sistemi intelligenti, 35(2):435–444.
- Peirce, C. S. (1935). Collected papers of charles sanders peirce. vol. v, Pragmatism and Pragmaticism.
- Lectures on Pragmatism. Meiner.
- Reasoning with language model prompting: A survey. ArXiv, abs/2212.09597.
- Nonlinear dimensionality reduction by locally linear embedding. science, 290(5500):2323–2326.
- ScriptByAI (2023). The complete list of ChatGPT plugins in ChatGPT plugin store. Accessed on December 25, 2023.
- StackExchange (2013). Converting a distance matrix into euclidean vector. Accessed on December 25, 2023.
- Challenging big-bench tasks and whether chain-of-thought can solve them. arXiv preprint arXiv:2210.09261.
- Torgerson, W. S. (1952). Multidimensional scaling: I. theory and method. Psychometrika, 17(4):401–419.
- The importance of being recurrent for modeling hierarchical structure. ArXiv, abs/1803.03585.
- Emergent abilities of large language models. arXiv preprint arXiv:2206.07682.
- Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
- When A.I. chatbots hallucinate. The New York Times. Available at: https://www.nytimes.com/2023/05/01/business/ai-chatbots-hallucination.html. Accessed on January 3, 2024.
- A fast approximation to multidimensional scaling. In IEEE workshop on computation intensive methods for computer vision.
- Mark Glickman (2 papers)
- Yi Zhang (994 papers)