Papers
Topics
Authors
Recent
Search
2000 character limit reached

Deterministic Dynamic Execution

Updated 4 February 2026
  • Deterministic Dynamic Execution is a scheduling paradigm that enforces pre-determined resource allocation and execution order to guarantee timing-anomaly-free execution of multi-typed DAG tasks.
  • The approach uses an offline simulation with WCET assignments to derive both safe and tight worst-case response time estimates under deterministic constraints.
  • Empirical evaluations reveal significant reductions in worst-case response times and jitter with minimal scheduling overhead on heterogeneous systems.

Deterministic Dynamic Execution (DDE) is a scheduling paradigm introduced to guarantee timing-anomaly-free execution for dynamic scheduling of multi-typed directed acyclic graph (DAG) tasks on heterogeneous platforms. By enforcing deterministic constraints on resource allocation and task execution order, DDE eliminates the pitfalls of conventional dynamic schedulers, notably timing anomalies that complicate or invalidate standard worst-case response time (WCRT) analyses. DDE achieves both safety and tightness in WCRT estimation through a single offline simulation, with minimal overhead and favorable empirical properties even on large-scale heterogeneous systems (Zhu et al., 28 Jan 2026).

1. System and Task Model

DDE is formulated for heterogeneous platforms comprising multiple processing unit types (e.g., CPUs, GPUs), each possibly with several identical instances. Let PUT\mathrm{PUT} denote the set of processing-unit types, and resNum(res)\mathrm{resNum}(res) the instance count for type resPUTres \in \mathrm{PUT}.

The task model consists of a multi-typed DAG G=(V,E,M,Γ)G=(V,E,M,\Gamma):

  • VV: tasks; EV×VE \subseteq V \times V: precedence constraints.
  • M:V2PUTM: V \to 2^{\mathrm{PUT}} \setminus \emptyset: eligible resource types for each task.
  • Γ:V×PUTI>0×I>0\Gamma: V \times \mathrm{PUT} \to \mathbb{I}_{>0} \times \mathbb{I}_{>0} giving for (v,res)(v, res)

Γ(v,res)=[BCET(v,res), WCET(v,res)]\Gamma(v, res) = [\mathrm{BCET}(v, res),\ \mathrm{WCET}(v, res)]

where BCET/WCET denote best/worst-case execution times, capturing workload- and resource-dependent variability.

For task viv_i:

  • sis_i: start time; fif_i: finish time; RT(i)=fissrcRT(i) = f_i - s_\mathrm{src} (relative to the DAG source).
  • System response: RT(vsink)RT(v_\mathrm{sink}).

Let S\mathbb{S} be the set of all feasible schedules respecting precedence and resource constraints. Then the system WCRT is:

WCRT=maxSSRTS(vsink)RT(vsink)when all tasks use WCETs\mathrm{WCRT} = \max_{\mathcal{S} \in \mathbb{S}} RT_\mathcal{S}(v_\mathrm{sink}) \approx RT(v_\mathrm{sink}) \,\, \text{when all tasks use WCETs}

2. Timing Anomalies in Dynamic Scheduling

A timing anomaly (TA) refers to the counterintuitive phenomenon where a locally accelerated event (e.g., a task finishes early at BCET) can increase the global WCRT relative to the all-WCET simulation bound. This undermines the validity of conventional WCRT analyses and frequently forces unsafe or overly pessimistic estimation procedures.

The paper defines a strict timing anomaly precisely. Consider two system states c1,c2c_1, c_2 in the execution-progress domain (see below), with c1(t)Pc2(t)c_1(t) \sqsubset_P c_2(t) for all tasks tt in some ST\mathbb{S} \subseteq \mathbb{T}, and for all future evolution, allocated execution times in c1c_1 are at least those in c2c_2. A strict timing anomaly is exhibited if, after nn transitions, updn(c1)̸Cupdn(c2)upd^n(c_1) \not\sqsubseteq_C upd^n(c_2), i.e., the “faster” state falls behind due to dynamically shifting resource contention or precedence scheduling.

An explicit example demonstrates this: an early completion on a GPU causes downstream loss of parallelism, making a critical task await resource availability longer, prolonging the overall response beyond the offline bound (Zhu et al., 28 Jan 2026).

3. Deterministic Dynamic Execution Algorithm

3.1 Execution-Progress Model

Execution progress is represented by the domain

S={Block,Ready,Exec,Finish}S = \{\mathrm{Block},\, \mathrm{Ready},\, \mathrm{Exec},\, \mathrm{Finish}\}

and P=S×N0P = S \times \mathbb{N}_0, with p=(s,t)p = (s, t) indicating the current stage and the remaining time at that stage. A partial order P\sqsubseteq_P is imposed:

(s1,t1)P(s2,t2)    s1Ss2or(s1=s2t1t2)(s_1, t_1) \sqsubseteq_P (s_2, t_2) \iff s_1 \sqsubseteq_S s_2 \quad \text{or} \quad (s_1 = s_2 \land t_1 \ge t_2)

where BlockSReadySExecSFinish\mathrm{Block} \sqsubset_S \mathrm{Ready} \sqsubset_S \mathrm{Exec} \sqsubset_S \mathrm{Finish}.

A system state c:TPc: \mathbb{T} \to P maps each task tt to its progress, ordered pointwise: cCcc \sqsubseteq_C c' iff c(t)Pc(t)c(t) \sqsubseteq_P c'(t) for all tt.

The state transition upd:CCupd: C \to C is defined such that in each scheduling tick, for each task tt:

  • If ready to move stage (all dependencies met and resources available) or Exec done: transition to next stage (stage(t)stage'(t)) and set time to required allocation, otherwise decrement remaining time.

3.2 Deterministic Execution Constraints

To exclude timing anomalies, DDE enforces:

  • Resource Allocation Determinism: Every task tt is always scheduled on a predetermined unit type restres_t:

resAllocPatD={(t,rest)tT}resAllocPat_D = \{(t, res_t)\mid t \in \mathbb{T}\}

  • Execution Order Determinism: Tasks must enter Exec in a single total order OD=(t1t2tN)O_D = (t_1 \preceq t_2 \preceq \cdots \preceq t_N) extending the DAG’s partial order. Task tjt_j cannot start until all titjt_i \prec t_j have started.

The DDE scheduler executes as follows (pseudocode as stated):

1
2
3
4
5
6
7
while exists tick(c)==0:
    c  upd(c)
for each t in ReadyQueue in order O_D:
    if all predecessors of t in O_D have begun Exec:
        assign t to an idle instance of res_t
        set its remaining time = WCET(t, res_t)
        mark t as begun Exec
Complexity per tick is O(ReadyQueue+logN)O(|\mathrm{ReadyQueue}| + \log N), assuming ODO_D is maintained via a priority queue.

4. WCRT Estimation via Offline Simulation

Given the enforced determinism, the WCRT is derived from a single offline execution:

  • Assign all tasks their WCETs and simulate updupd from all tasks in Block.
  • Upon sink completion, response time RT(vsink)RT(v_\mathrm{sink}) is both safe (no under-approximation) and tight (no excessive pessimism), as deterministic constraints preclude timing anomaly deviation.

Formally,

WCRT=minresAllocPatD,ODmaxn(fsink(n))\mathrm{WCRT} = \min_{\substack{resAllocPat_D,\, O_D}} \max_n \left( f_{\mathrm{sink}}(n) \right)

where fsink(n)f_{\mathrm{sink}}(n) is the finish time across all ticks.

5. Construction of Deterministic Execution Constraints

DDE constraints can be generated offline by either of two methods:

  • Trace Extraction: Run any dynamic scheduler (e.g., HBFS) offline under all-WCET settings. For each task, record its start time, finish time, and chosen resource. Build ODO_D by sorting tasks by sts_t (ties broken by task ID), set resAllocPatD(t)=restresAllocPat_D(t)=res_t.
  • Heuristic-Based HACPA: Compute for each task a critical-path “rank”:

rank(t)=avgCost(t)+maxusucc(t)rank(u)\mathit{rank}(t) = \mathrm{avgCost}(t) + \max_{u \in succ(t)} \mathit{rank}(u)

with avgCost(t)=1M(t)resM(t)WCET(t,res)\mathrm{avgCost}(t) = \frac{1}{|M(t)|} \sum_{res \in M(t)} \mathrm{WCET}(t, res). Sort tasks by descending rank. For each tt, choose the resource leading to earliest finish, update ODO_D and resAllocPatDresAllocPat_D in assignment order.

6. Formal Guarantees and Correctness

The DDE paradigm is supported by a series of formal results:

  • Monotonicity: Under DDE, execution progress is monotonic—no fast-execution state can ever fall behind a slower one under any evolution.
  • Strict TA Elimination: Lemmas ensure forward progress, dependency monotonicity, and that staged transitions preserve the progress order. The general strict TA-freedom theorem constructs a general proof by induction and repeated application of monotonicity lemmas, showing that no dynamic execution under DDE can exhibit a strict timing anomaly.

This ensures the single all-WCET simulation bound is both sound (never optimistic) and maximally precise given the enforced constraints.

7. Empirical Evaluation and Observed Impact

Experiments employed random DAG task sets generated via the G(n,p)G(n, p) model, with n{10,,40}n \in \{10, \dots, 40\} and edge-probability p{0.1,,0.9}p \in \{0.1, \dots, 0.9\}. Variability in Γ(v,res)\Gamma(v, res) was chosen to represent both wide and narrow WCET-BCET spreads.

Key findings:

  • TA Probability: In unconstrained HFCFS, TAs occurred in up to 44.8% of tested systems at low pp, vanishing as DAG density increased. HBFS displayed much lower rates (2.6%)\leq 2.6\%).
  • WCRT Reduction: In TA-affected systems, DDE imposed on HFCFS reduced WCRT by between 5–25% (best observed: 72%) relative to the unconstrained maximum observed WCRT. HACPA-derived constraints further reduced WCRT.
  • Jitter: Enforced determinism cut response-time jitter by 7–9% on average.
  • Average Response Time Impact: The average overhead on AVRT was minor (1–12%; avg 4.2%). In TA-affected systems, AVRT sometimes decreased (average –1.2%, up to –40%).
  • Overhead: Scheduling overhead for DDE was minimal (one priority-queue lookup per ready task per tick) relative to task execution costs.

A plausible implication is that the robustness and tight WCRT estimation provided by DDE are achieved with negligible runtime cost and limited sacrifice in average-case performance, supporting its practical suitability for real-world safety- or real-time-critical heterogeneous applications (Zhu et al., 28 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Deterministic Dynamic Execution.