Procedural Text Mining with Large Language Models (2310.03376v1)
Abstract: Recent advancements in the field of Natural Language Processing, particularly the development of large-scale LLMs that are pretrained on vast amounts of knowledge, are creating novel opportunities within the realm of Knowledge Engineering. In this paper, we investigate the usage of LLMs in both zero-shot and in-context learning settings to tackle the problem of extracting procedures from unstructured PDF text in an incremental question-answering fashion. In particular, we leverage the current state-of-the-art GPT-4 (Generative Pre-trained Transformer 4) model, accompanied by two variations of in-context learning that involve an ontology with definitions of procedures and steps and a limited number of samples of few-shot learning. The findings highlight both the promise of this approach and the value of the in-context learning customisations. These modifications have the potential to significantly address the challenge of obtaining sufficient training data, a hurdle often encountered in deep learning-based Natural Language Processing techniques for procedure extraction.
- Process Extraction from Text: state of the art and challenges for the future. CoRR abs/2110.03754 (2021).
- Leveraging pre-trained language models for conversational information seeking from text. arXiv preprint arXiv:2204.03542 (2022).
- Semantic modeling and analysis of complex data-aware processes and their executions. Expert Syst. Appl. 198 (2022), 116702.
- Transformation from human-readable documents and archives in arc welding domain to machine-interpretable data. Comput. Ind. 128 (2021), 103439.
- Daniel Garijo and Yolanda Gil. 2012. Augmenting prov with plans in p-plan: scientific processes as linked data. CEUR Workshop Proceedings.
- Better Call the Plumber: Orchestrating Dynamic Information Extraction Pipelines. In ICWE (Lecture Notes in Computer Science, Vol. 12706). Springer, 240–254.
- Dissociating language and thought in large language models: a cognitive perspective. arXiv:2301.06627 [cs.CL]
- Recent Advances in Natural Language Processing via Large Pre-Trained Language Models: A Survey. ACM Comput. Surv. (jun 2023). https://doi.org/10.1145/3605943
- Unifying Large Language Models and Knowledge Graphs: A Roadmap. arXiv:2306.08302 [cs.CL]
- Adrian Rebmann and Han van der Aa. 2021. Extracting Semantic Process Information from the Natural Language in Event Logs. In CAiSE, Vol. 12751. Springer, 57–74.
- K-Hub: A Modular Ontology to Support Document Retrieval and Knowledge Extraction in Industry 5.0. In ESWC 2023 Proceedings (LNCS, Vol. 13870). Springer, 454–470.
- Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. arXiv preprint arXiv:2206.04615 (2022).
- Head-to-Tail: How Knowledgeable are Large Language Models (LLM)? A.K.A. Will LLMs Replace Knowledge Graphs? arXiv:2308.10168 [cs.CL]
- Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022).