Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 30 tok/s
Gemini 2.5 Pro 46 tok/s Pro
GPT-5 Medium 18 tok/s Pro
GPT-5 High 12 tok/s Pro
GPT-4o 91 tok/s Pro
Kimi K2 184 tok/s Pro
GPT OSS 120B 462 tok/s Pro
Claude Sonnet 4 36 tok/s Pro
2000 character limit reached

AI-Assisted Modeling: DSL-Driven AI Interactions (2509.05160v1)

Published 5 Sep 2025 in cs.PL and cs.SE

Abstract: AI-assisted programming greatly increases software development performance. We enhance this potential by integrating transparency through domain-specific modeling techniques and providing instantaneous, graphical visualizations that accurately represent the semantics of AI-generated code. This approach facilitates visual inspection and formal verification, such as model checking. Formal models can be developed using programming, natural language prompts, voice commands, and stage-wise refinement, with immediate feedback after each transformation step. This support can be tailored to specific domains or intended purposes, improving both code generation and subsequent validation processes. To demonstrate the effectiveness of this approach, we have developed a prototype as a Visual Studio Code extension for the Lingua Franca language. This prototype showcases the potential for novel domain-specific modeling practices, offering an advancement in how models are created, visualized, and verified.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

  • The paper introduces a novel DSL-driven AI modeling framework that enhances transparency and iterative refinement with multi-modal input and real-time visualization.
  • The methodology integrates state-of-the-art LLM-based synthesis with domain-specific IDEs, leveraging speech, text, and code editing for immediate feedback.
  • The prototype demonstrates low-latency interactions and improved error correction using automated Tool API functions and a visual feedback loop.

AI-Assisted Modeling: DSL-Driven AI Interactions

Introduction and Motivation

The paper "AI-Assisted Modeling: DSL-Driven AI Interactions" (2509.05160) addresses the integration of AI—specifically LLMs—into domain-specific modeling workflows, with a focus on enhancing transparency, controllability, and iterative refinement. The authors identify the limitations of current AI-assisted programming tools, which are predominantly tailored to general-purpose languages and often rely on one-shot prompt paradigms. These approaches are insufficient for domain-specific languages (DSLs), especially those with complex or underrepresented syntax, such as Lingua Franca or SCCharts. The paper proposes a hybrid methodology that combines state-of-the-art AI-assisted IDEs with domain-specific modeling techniques, emphasizing instantaneous graphical visualizations and fine-grained observation and interaction points throughout the modeling process.

Conceptual Framework: Iterative, Multi-Modal Modeling

The core contribution is a workflow that enables modelers to interact with the system via speech, text, or direct code editing, receiving immediate feedback at multiple transformation stages. This is illustrated in the modeling concept diagram, which formalizes the feedback loop between the modeler and the system. Figure 1

Figure 1: The modeling concept of DSL-driven AI interaction: A modeler interacts with the system through speech or text inputs and receives feedback from multiple transformation stages, including instantaneous diagram synthesis and compilation results. This workflow forms a feedback loop, enabling both the modeler and the system to incrementally refine the model.

The workflow is characterized by:

  • Multi-modal input: Speech, natural language, and direct code editing are all supported as entry points.
  • Incremental transformation: Each input is processed through a series of transformation stages, with observation points (OPs) and interaction points (IPs) at each stage.
  • Immediate visualization: Textual models are automatically synthesized into semantically accurate diagrams, facilitating rapid inspection and correction.
  • System-level validation: Automata learning and model checking are integrated for formal verification.

This approach is particularly effective for DSLs with graphical representations, where visual feedback is essential for understanding complex relationships and semantics.

Prototype Implementation: Visual Studio Code Extension

The authors present a prototype implemented as a Visual Studio Code extension for the Lingua Franca language, targeting cyber-physical systems. The extension leverages the OpenAI Tool API for context-aware code generation and integrates the KIELER diagram engine for automatic diagram synthesis. The workflow is demonstrated in the editor, where the modeler can interact via voice, text, or direct editing, and observe the outputs at each stage. Figure 2

Figure 2: Demonstrator of rapid, natural modeling with Lingua Franca in Visual Studio Code. The observation points display the outputs of the speech transcription, the generated textual model, and the synthesized diagram, respectively. The interaction points allow the modeler to provide input via speech, text prompt, or direct editing of the model.

The prototype supports:

  • Speech-to-text transcription (using Whisper-1 or similar models)
  • LLM-based model synthesis (using o4-mini or interchangeable LLMs)
  • Automatic diagram generation (via KIELER/ELK)
  • Simultaneous observation and interaction at all stages

The demonstrator showcases the creation of a Lingua Franca model in under two minutes, with all prompts spoken, transcribed, and synthesized into both textual and graphical representations. Figure 3

Figure 3: Demonstration of the prototype in action, following the transcript from the accompanying video. The time progresses from top to bottom and the model was created in less than 2 min. All prompts on the left side were spoken by the modeler, transcribed, and then sent to the LLM to generate the Lingua Franca source model, which is then synthesized into the corresponding diagram. The final Lingua Franca model source is shown to the right.

Technical Details: Tool API and Grammar Integration

A key technical innovation is the use of the Tool API to expose DSL grammar constructs to the LLM. Rather than relying solely on prompt engineering, the Tool API provides structured, grammar-aligned functions that the LLM can invoke to generate syntactically correct code fragments. This approach offloads contextualization from the prompt to the tool layer, improving reliability and scalability, especially for DSLs with complex or polyglot syntax.

The authors demonstrate that Tool API definitions can be generated automatically from annotated grammar files, reducing manual effort and ensuring consistency between the DSL specification and the AI assistant's capabilities. This meta-tooling approach is extensible to other DSLs and can be further automated by associating formal comments with grammar parameters.

Evaluation and Discussion

The prototype demonstrates fluent, low-latency interaction, with speech-to-text and LLM response times suitable for real-time modeling. The integration of observation and interaction points enables rapid correction and iterative refinement, addressing the shortcomings of black-box, one-shot prompt engineering. The graphical feedback loop is particularly effective for non-experts and non-native speakers, lowering the barrier to entry for DSL modeling.

The authors note that the optimal granularity of Tool API functions and the combination of retrieval, fine-tuning, and prompt engineering techniques remain open research questions. The approach is orthogonal to improvements in LLM architectures and can benefit from advances in retrieval-augmented generation and model fine-tuning.

Implications and Future Directions

The proposed methodology has significant implications for both practical tooling and theoretical modeling workflows:

  • Enhanced controllability: Fine-grained observation and interaction points enable modelers to guide the AI assistant more effectively, improving reliability and reducing the risk of semantic errors.
  • Transparency and verification: Instantaneous graphical synthesis and system-level validation facilitate formal verification and visual inspection, critical for safety-critical and cyber-physical systems.
  • Meta-tooling automation: The ability to generate Tool API definitions from DSL grammars paves the way for self-adaptive modeling environments, where the modeling tools themselves can be partially synthesized by AI.
  • Generalizability: While demonstrated for Lingua Franca, the approach is applicable to other DSLs, particularly those with graphical semantics or modular architectures.

Future research directions include optimizing the granularity of Tool API functions, integrating richer diagnostic feedback (e.g., compiler errors, model checking results) into the feedback loop, and extending the methodology to support collaborative, multi-agent modeling scenarios.

Conclusion

This work presents a comprehensive framework for AI-assisted modeling in DSLs, combining LLM-based synthesis, multi-modal interaction, and instantaneous graphical feedback. The prototype implementation in Visual Studio Code for Lingua Franca demonstrates the feasibility and advantages of the approach, particularly in terms of iterative refinement, transparency, and accessibility. The methodology is extensible, tool-agnostic, and compatible with ongoing advances in LLMs and AI-assisted development environments. The integration of meta-tooling and formal grammar alignment represents a promising direction for the future of intelligent, adaptive modeling systems.

Youtube Logo Streamline Icon: https://streamlinehq.com

alphaXiv