Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Procedural Text Mining with Large Language Models (2310.03376v1)

Published 5 Oct 2023 in cs.CL, cs.AI, cs.IT, and math.IT

Abstract: Recent advancements in the field of Natural Language Processing, particularly the development of large-scale LLMs that are pretrained on vast amounts of knowledge, are creating novel opportunities within the realm of Knowledge Engineering. In this paper, we investigate the usage of LLMs in both zero-shot and in-context learning settings to tackle the problem of extracting procedures from unstructured PDF text in an incremental question-answering fashion. In particular, we leverage the current state-of-the-art GPT-4 (Generative Pre-trained Transformer 4) model, accompanied by two variations of in-context learning that involve an ontology with definitions of procedures and steps and a limited number of samples of few-shot learning. The findings highlight both the promise of this approach and the value of the in-context learning customisations. These modifications have the potential to significantly address the challenge of obtaining sufficient training data, a hurdle often encountered in deep learning-based Natural Language Processing techniques for procedure extraction.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (14)
  1. Process Extraction from Text: state of the art and challenges for the future. CoRR abs/2110.03754 (2021).
  2. Leveraging pre-trained language models for conversational information seeking from text. arXiv preprint arXiv:2204.03542 (2022).
  3. Semantic modeling and analysis of complex data-aware processes and their executions. Expert Syst. Appl. 198 (2022), 116702.
  4. Transformation from human-readable documents and archives in arc welding domain to machine-interpretable data. Comput. Ind. 128 (2021), 103439.
  5. Daniel Garijo and Yolanda Gil. 2012. Augmenting prov with plans in p-plan: scientific processes as linked data. CEUR Workshop Proceedings.
  6. Better Call the Plumber: Orchestrating Dynamic Information Extraction Pipelines. In ICWE (Lecture Notes in Computer Science, Vol. 12706). Springer, 240–254.
  7. Dissociating language and thought in large language models: a cognitive perspective. arXiv:2301.06627 [cs.CL]
  8. Recent Advances in Natural Language Processing via Large Pre-Trained Language Models: A Survey. ACM Comput. Surv. (jun 2023). https://doi.org/10.1145/3605943
  9. Unifying Large Language Models and Knowledge Graphs: A Roadmap. arXiv:2306.08302 [cs.CL]
  10. Adrian Rebmann and Han van der Aa. 2021. Extracting Semantic Process Information from the Natural Language in Event Logs. In CAiSE, Vol. 12751. Springer, 57–74.
  11. K-Hub: A Modular Ontology to Support Document Retrieval and Knowledge Extraction in Industry 5.0. In ESWC 2023 Proceedings (LNCS, Vol. 13870). Springer, 454–470.
  12. Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. arXiv preprint arXiv:2206.04615 (2022).
  13. Head-to-Tail: How Knowledgeable are Large Language Models (LLM)? A.K.A. Will LLMs Replace Knowledge Graphs? arXiv:2308.10168 [cs.CL]
  14. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022).
Citations (4)

Summary

We haven't generated a summary for this paper yet.