Controlled SANs for Real-Time Systems
- Controlled SANs are formal extensions of classical Stochastic Activity Networks that integrate explicit control actions with probabilistic and timed behaviors.
- They support both qualitative and quantitative analysis through hierarchical automata-theoretic semantics, bisimulation, and compositional design techniques.
- By generalizing CTMDPs and enabling robust model abstraction, Controlled SANs provide a rigorous framework for designing and verifying distributed real-time systems under uncertainty.
Controlled Stochastic Activity Networks (Controlled SANs) are a formal extension of classical Stochastic Activity Networks developed to systematically model distributed real-time systems exhibiting dynamic nondeterminism, probabilistic branching, and stochastic timing, with direct integration of explicit control actions. This framework introduces a rigorous mathematical and semantic foundation supporting compositional system design, model abstraction, and the precise specification, verification, and synthesis of dependable systems operating under uncertainty. Controlled SANs unify graphical notation, hierarchical automata-theoretic semantics, and a structured taxonomy of control policies, thus supporting both qualitative and quantitative analyses and generalizing legacy models such as continuous-time Markov decision processes (CTMDPs) (Movaghar, 17 Nov 2025).
1. Formal Specification of Controlled SANs
A Controlled SAN is formally defined as a 9-tuple: with the following components:
- : finite set of places
- : finite set of instantaneous activities
- : finite set of timed activities
- : finite set of control actions
- : finite set of input gates; each gate with arity has an enabling predicate and a state-transform function
- : finite set of output gates; each gate has a function
- : set of input relations, mapping places, ports, gates, and activities
- : set of instantaneous output relations
- : set of timed output relations, indexed by timed activities, control actions, output gates, ports, and places
A marking denotes the state configuration. Activities are enabled in a marking according to the evaluation of each input gate’s enabling predicate on the relevant place values. The completion of activities (instantaneous or timed) triggers input transformations followed by output transformations, which for timed activities are governed by a control action selected under a policy.
2. Hierarchical Automata-Theoretic Semantics
Controlled SANs admit a layered automata-theoretic interpretation via stochastic labeled transition systems: where:
- is the set of stable markings reachable from the initial marking , with an added sink state for Zeno behavior
- : timed activities
- : control-action alphabet
- : nondeterministic transitions
- : probabilistic transitions
- : timed transitions, encoding rates or duration distributions
- : initial stable states
For exponential-timed Controlled SANs, the framework maps precisely to CTMDPs through the transition rate function
and corresponding jump distribution
Key rules allow encoding instantaneous probabilistic choice, timed nondeterministic steps under policy, and stochastic timing.
3. Taxonomy and Expressiveness of Control Policies
Control policies in Controlled SANs range by history and computational memory:
- Memoryless policies:
- Finite-memory policies: Embedded automaton manages memory state evolution and control selection
- Stack-augmented policies: Use a finite stack () for higher expressive power
- Tape-augmented policies: (Turing-complete) maintain a read/write tape
Any history-dependent policy is a function
Class acceptance is organized into a strict hierarchy mirroring the Chomsky classification: These hierarchies are formalized for both local and global language classes, with each layer strictly more expressive than its predecessor.
4. Behavioral Equivalences and Isomorphism
Several forms of behavioral equivalence facilitate model abstraction and correctness analysis:
- Bisimulation (): A relation on state spaces of controlled automata ensures mutual simulation of transitions under matched control actions and activities. Isomorphism is characterized by bijective bisimulation.
- Probabilistic bisimulation (): Ensures preservation not only of transitions but also of transition probabilities, generalized for stochastic automata by requiring equality of rates and distributions.
These equivalences support compositional system reasoning, equivalence checking, and semantic-preserving abstraction for scalable analyses.
5. Relationship to Classical CTMDPs
Controlled SANs strictly generalize CTMDPs. When all timed activities are exponentially distributed and control policies resolve nondeterminism, the Markovian controlled stochastic automaton corresponding to a Controlled SAN can be encoded as a CTMDP via:
This strictly encompassing relationship allows Controlled SANs to subsume classical MDP and CTMDP synthesis and analysis approaches.
6. Abstraction Mechanisms and Compositional Reasoning
Controlled SANs support structured abstraction and composition:
- Bisimulation quotienting and GSMP/SMDP unfolding enable mapping complex models to simpler representations with preserved operational semantics.
- Composition operators like disjoint union () and synchronous product () provide closure properties for language families, supporting modular system design.
- Algebraic laws establish relationships among policy-driven languages, permitting set-theoretic manipulations (union, intersection, complement) within deterministic and memoryless fragments.
7. Analysis Techniques and Applications
Controlled SANs enable both qualitative and quantitative analysis:
- Qualitative verification (reachability/emptiness): Decidable in PTIME for memoryless/finite-memory policies; pushdown algorithms for stack-augmented; undecidable for tape/general history-driven policies.
- Quantitative analysis: MDP abstraction and value/policy iteration facilitate reward optimization; continuous-time synthesis (expected/discounted total-reward) executable via linear programming or fixed-point iteration. Time-bounded reachability is PSPACE-complete, sometimes requiring richer policy classes.
- General Controlled SANs (non-exponential): Embedding into generalized semi-Markov processes (GSMPs) and discretization enables computation of -optimal policies via contraction mappings and value iteration.
- PAC-optimality: Reinforcement learning algorithms yield, with high probability, -optimal policies for finite-state MDPs, with guarantees extending to CTMDPs and Controlled SANs by uniformization and discretization.
Controlled SANs therefore provide a compositional, policy-sensitive, and semantically layered framework for modeling and analyzing distributed real-time systems under uncertainty, enabling rigorous design and verification of safety-critical systems where control, timing, and stochasticity are tightly coupled (Movaghar, 17 Nov 2025).