Papers
Topics
Authors
Recent
Search
2000 character limit reached

Guidance Engineering

Updated 13 February 2026
  • Guidance Engineering is the systematic development and evaluation of mechanisms that enforce proper tool usage and workflow adherence.
  • It integrates methods from machine learning, robotics, and software engineering to ensure error detection and best practice communication.
  • Applications include API compliance, error correction, and validation metrics, enhancing reproducibility and reliability in complex systems.

Guidance engineering encompasses the systematic development, integration, and evaluation of mechanisms (languages, APIs, tools, workflows) that facilitate or enforce the correct use of computational tools, adherence to prescribed processes, and the achievement of domain-specific objectives. The concept spans multiple disciplines, including applied ML, control and robotics, reinforcement learning, prompt engineering for LLM-based extraction, and physical task instruction, each of which imposes unique technical requirements, artifacts, and reliability criteria.

1. Definition and Scope of Guidance Engineering

The foundational definition, as articulated by Reimann and Kniesel-Wünsche, posits guidance as the set of language, API, tool-level, or workflow-level mechanisms that “facilitate or enforce correct usage of a tool or an API or proper adherence to a workflow” (Reimann et al., 2022). It serves two primary functions: (1) prevention, detection, explanation, and correction of erroneous usage; and (2) communication of best practices and assistance in their application.

In complex technical systems, guidance mechanisms may address coding errors (API contract enforcement), empirical experimentation protocols (data/model handling in ML), optimal geometric behavior (flight guidance laws), or pedagogic/interactive feedback (physical skill training). The domain-specific artifacts subject to guidance include code, data, trained models, evaluation metrics, trajectories, and user actions.

2. Taxonomies and Classification Frameworks

Topical taxonomies in guidance engineering are inherently discipline-dependent. In ML engineering, the spectrum of guidance needs is mapped to the canonical supervised-learning workflow, subdivided as follows (Reimann et al., 2022):

Phase Principal Guidance Requirements
Evaluation-Focused RE Metric selection, benchmarking, requirements specification
Test-Driven ML Development Data splitting strategies, test-set “sealing,” stratification
Data Engineering Missing value imputation, categorical encoding, scaling, feature relevance, pipeline propagation
Model Engineering Algorithm selection, API/hyperparameter validity, error messaging
Model Quality Engineering Experiment tracking, overfitting/underfitting detection, reproducibility

In modeling guidance for software engineering, the relevant dimensions are classified as (Chakraborty et al., 2022):

Dimension Example Instantiations
Focus “What” (which elements), “How” (workflow steps), or both
Views/Perspectives Multi-view (e.g., correctness, clarity), or single-view
Decomposition Hierarchical/incremental vs. flat guidance
Practices/Anti-patterns Explicit “do’s/don’ts” or none
Cognition User style/mental model consideration

This multi-dimensional structure enables fine-grained mapping of guidance artifacts, facilitating comparative analysis and tool-chain design.

3. Mechanisms and Best Practices for Guidance Realization

Guidance mechanisms manifest as a synergy of static (compile-time) and dynamic (run-time) techniques, formal specifications, and user-facing feedback components.

ML Application Development (Reimann et al., 2022):

  • Static/lint checks to enforce dataset partitioning and test-set integrity.
  • Precondition enforcement on API inputs (e.g., parameter constraints, type checks).
  • Dynamic best practice flags (e.g., regularization suggestions for overfitting).
  • Enhanced error messages with specific corrective advice.

Modeling in Software Engineering (Chakraborty et al., 2022):

  • Explicit documentation of methodology, style, and guideline distinctions.
  • Checklists or IDE-integrated wizards that implement methodology or enforce guidelines.
  • Provision of anti-pattern detectors and workflow orchestrators.
  • User profiling and adaptive interfaces for style-cognizant assistance.

Prompt Engineering for Extraction (Khatami et al., 2024):

  • Modular, schema-scaffolded prompts with strict output contracts (e.g., JSON).
  • Role- and context-specific instructions to prevent hallucinations and ensure specification completeness.
  • Iterative refinement and validation steps, including machine validation and spot annotation.

Physical Task Guidance with AI/MR (Caetano et al., 2024):

  • Encapsulation of interaction design into “Design Considerations” (teaching vs. directing, timing, error handling).
  • Taxonomy of Design Patterns and an Interaction Canvas to analyze user-system-execution/evaluation gulfs.
  • Multimodal and adaptive feedback channels (MR overlays, haptics, chain-of-thought explanations).

4. Formal Models: Contracts, Constraints, and Quantitative Guidance

Guidance effectiveness and expressivity are linked to formal models of constraints and verification. In ML automation (Reimann et al., 2022), four classes of API constraints are defined:

  • Type: parameter input types (e.g., kernel ∈ String ∪ Callable).
  • Dependency: parameter interrelationships (e.g., degree valid only if kernel="poly").
  • Temporal: sequencing/precondition requirements.
  • Execution context: resource constraints (e.g., verbosity under multithreading).

In model extraction and process compliance, these constraints are encoded in domain-specific languages (DSLs), static analysis frameworks, and metadata ontologies, supporting compile-time detection of misconfiguration and semantic violations.

Empirical effectiveness is evaluated by metrics such as:

  • Pass-rate improvement (e.g., pass@1 on code repair tasks in Agent-RLVR: 9.4% → 22.4% with guidance (Da et al., 13 Jun 2025)).
  • Capability-compliance rates to predict OOD performance in ML models (Yang et al., 2022).
  • Coverage increases and schema conformance in prompt-guided extraction (>90% field coverage in well-structured prompts (Khatami et al., 2024)).

5. Gaps, Challenges, and Proposed Extensions

Empirical surveys and technical critiques highlight pervasive shortfalls (Reimann et al., 2022, Chakraborty et al., 2022):

  • Inconsistent or opaque API documentation, hiding critical constraints in natural language.
  • Weak or nonexistent static checking of best practices/hyperparameters; late surfacing of errors.
  • Generic, non-actionable error messages.
  • Absence of workflow-level enforcement for empirically critical routines (stratified splits, pipeline replay).
  • Minimal attention to stakeholder/tooling concerns and end-user cognitive variability.
  • Sparse empirical validation of guidance methods, especially in production settings.

Recommended extensions include:

  • Unified, statically-typed front-end APIs with domain-specific pre—/post—conditions and framework-wide enforcement of best practices.
  • DSLs for ML and modeling workflows that encode all four constraint types and elevate guidance errors to compile-time events.
  • Ontology-driven suggestions and metadata tracking for data, model, and provenance semantics.
  • Semantic IDE integration for live static analysis, experiment management, and quickfixes.

6. Research Directions and Open Opportunities

Outstanding research challenges identified include (Reimann et al., 2022):

  • Development of contract/specification languages for API and pipeline constraints encompassing temporal, dependency, and context information.
  • Static analysis techniques for data pipelines, including missing-value propagation, drift detection, and feature-semantics verification.
  • Ontology-based recommendation systems to leverage historic experiment metadata (public and private benchmarks) for guidance on model/algorithm selection.
  • Experiment reproducibility frameworks capturing all relevant metadata (random seeds, hardware, pipeline versions) for exact reruns.
  • Controlled studies of productivity and learning-curve improvements enabled by guidance mechanisms.
  • Adaptation and personalization of guidance to account for cognitive styles, user profiles, or domain expertise.

7. Synthesis and Impact Across Domains

Guidance engineering is an interdisciplinary practice, critical both for ensuring technical correctness (in ML, control, and software modeling) and for scaffolding human understanding and skill acquisition (in prompt engineering and physical task guidance). By formalizing workflows, encoding best practices as explicit contracts, and enriching user interactions with actionable, context-aware feedback, guidance engineering enables reproducible, reliable, and efficient development in complex, error-prone computational domains. The maturity of guidance engineering is marked by the breadth of artifacts it can rigorously enforce, the integration of static and dynamic checks, and the degree to which it closes the empirical knowledge gap between domain experts and non-specialists (Reimann et al., 2022, Chakraborty et al., 2022, Khatami et al., 2024, Caetano et al., 2024, Yang et al., 2022).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Guidance Engineering.