Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 69 tok/s
Gemini 2.5 Pro 58 tok/s Pro
GPT-5 Medium 32 tok/s Pro
GPT-5 High 29 tok/s Pro
GPT-4o 108 tok/s Pro
Kimi K2 198 tok/s Pro
GPT OSS 120B 461 tok/s Pro
Claude Sonnet 4.5 33 tok/s Pro
2000 character limit reached

Assessing the Capability of Large Language Models for Domain-Specific Ontology Generation (2504.17402v1)

Published 24 Apr 2025 in cs.AI

Abstract: LLMs have shown significant potential for ontology engineering. However, it is still unclear to what extent they are applicable to the task of domain-specific ontology generation. In this study, we explore the application of LLMs for automated ontology generation and evaluate their performance across different domains. Specifically, we investigate the generalizability of two state-of-the-art LLMs, DeepSeek and o1-preview, both equipped with reasoning capabilities, by generating ontologies from a set of competency questions (CQs) and related user stories. Our experimental setup comprises six distinct domains carried out in existing ontology engineering projects and a total of 95 curated CQs designed to test the models' reasoning for ontology engineering. Our findings show that with both LLMs, the performance of the experiments is remarkably consistent across all domains, indicating that these methods are capable of generalizing ontology generation tasks irrespective of the domain. These results highlight the potential of LLM-based approaches in achieving scalable and domain-agnostic ontology construction and lay the groundwork for further research into enhancing automated reasoning and knowledge representation techniques.

Summary

Assessing the Capability of LLMs for Domain-Specific Ontology Generation

The paper entitled "Assessing the Capability of LLMs for Domain-Specific Ontology Generation" addresses an emerging aspect of artificial intelligence—the capacity of LLMs to facilitate ontology engineering, particularly in domain-specific contexts. Ontologies are crucial for structuring knowledge representation and achieving semantic interoperability in numerous fields, including healthcare and environmental sciences. Traditional methods of ontology engineering are notably labor-intensive and reliant on deep domain expertise. Thus, the automation potential offered by LLMs is of significant interest.

Methodology and Experimental Framework

The paper explores two state-of-the-art LLMs: DeepSeek and OpenAI's o1-preview, both claimed to possess reasoning capabilities. These models were tasked with generating ontological structures based on 95 curated competency questions (CQs) derived from six domains. CQs to ontological concepts were linked through user stories, providing contextual background for ontology generation. This setup seeks to rigorously test the efficacy of LLMs in interpreting prompts and translating them into coherent ontological drafts.

The experimental approach uses a novel dataset that spans topics such as Circular Economy, Music, Events, Microbe Habitat, Carbon and Nitrogen Cycling, and Water and Health. The dataset includes easy and hard questions, categorized to assess whether complexity affects LLM performance uniformly across domains. Results from this investigation aim to highlight the generalizable nature of LLM-driven ontology generation processes.

Results and Analysis

The findings emphasize that both models demonstrate high accuracy without significant variance across different domains of application. For instance, OpenAI's o1-preview encountered eight unmodeled CQs, while DeepSeek had five, indicating strong capabilities to produce domain-independent results. Minor modelling issues were identified but not considered critical, suggesting the models' robustness in handling more complex ontological generation tasks universally.

A notable result was the relatively consistent performance of both models across "easy" and "hard" competency questions. This indicates a robustness in their reasoning capabilities, challenging the assumption that greater complexity might degrade model efficacy. It also suggests their potential suitability for broader applications where competency question intricacy varies.

Implications and Future Considerations

The ability of LLMs to generate ontologies across varied domains carries significant implications. Practically, this represents a move toward scalable and more efficient ontology engineering processes, reducing the burden on domain experts in generating formal knowledge representations.

Theoretically, the results indicate a promising avenue for leveraging machine learning models with innate reasoning capacities in knowledge engineering tasks. Future research might focus on enhancing these reasoning capabilities further and could also explore adaptive learning approaches that refine ontology generation processes based on feedback.

Additionally, the application of LLMs could extend beyond ontology generation, potentially influencing wider knowledge graph engineering tasks, automatic concept extraction from natural language, and dynamic ontology alignment. As LLM capabilities continue to evolve, their integration into knowledge engineering frameworks could redefine best practices in semantic web technologies.

Conclusion

This paper underscores the emergent role of LLMs in transforming ontology engineering. By systematically evaluating their domain-agnostic capabilities, it reveals their potential as tools that significantly lower the barriers to efficient and scalable ontology generation. The paper's insights set the stage for future expansions where LLMs might increasingly assist or even autonomously execute various knowledge representation and reasoning tasks within artificial intelligence ecosystems.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube