SOOTT: Smoothed Online Optimization for Tracking
- SOOTT is a unified framework that integrates trajectory tracking, adversarial perturbation management, and smooth decision changes for real-time control.
- The BEST algorithm achieves robust tracking by emulating an ideal informed strategy without requiring direct adversarial inputs.
- The CoRT algorithm enhances performance by incorporating calibrated ML predictions, balancing consistency with robustness in dynamic environments.
Smoothed Online Optimization for Target Tracking (SOOTT) is a principled framework that unifies online trajectory tracking, robustness against adversarial perturbations, and control over decision smoothness via switching costs. SOOTT generalizes classical online convex optimization and control approaches by explicitly formulating the agent’s per-stage cost as a combination of tracking error, adversarial response, and penalization of abrupt decision changes. This formulation is well suited for applications such as real-time workload scheduling, adaptive resource allocation, and multi-target tracking, where simultaneous consideration of long-term performance, adversarial environments, and real-time adaptability is critical.
1. Core Problem Formulation and Cost Components
The SOOTT framework requires the agent to select actions at each time , incurring three distinct types of costs:
- Tracking Cost: Quantifies the deviation of a (typically windowed) average of the agent’s actions from a prescribed dynamic trajectory target .
where aggregates the previous actions.
- Adversarial Perturbation Cost: Penalizes discrepancies between the action and an adversary’s dynamic (and possibly hostile) target , captured by a convex function .
- Switching Cost: Penalizes abrupt changes in decisions to encourage smooth temporal evolution.
The cumulative cost per round is thus:
This formulation models, for instance, tracking a fluctuating service-level target () while meeting unpredictable real-time demand spikes (, e.g., inelastic jobs) and minimizing disruptive changes in resource allocation.
2. Robust Algorithmic Design: The BEST Algorithm
BEST (Backward Evaluation for Sequential Targeting) is a robust online algorithm for SOOTT that does not rely on knowledge of the adversarial target at decision time. Instead, BEST emulates the decisions of an idealized Informed Greedy Algorithm (IGA) (which knows ), by referencing IGA’s prior actions to drive its own update rule:
where accumulates IGA’s past decisions. The adversarial term is omitted since is not observable, making BEST robust to worst-case perturbations by design.
Theoretical guarantees are expressed using a degradation factor:
with
where and are problem-dependent parameters (e.g., the action dimension and window size). This performance bound ensures that even in adversarial environments, BEST incurs at most a constant-factor higher cost than the unattainable informed strategy.
3. Learning-Augmented Optimization: The CoRT Algorithm
CoRT (Consistent and Robust Tracking) extends BEST by incorporating untrusted predictions of the adversarial target , typically provided by black-box machine learning models (such as LSTMs). Rather than trusting predictions blindly, CoRT implements a calibration mechanism:
- Let denote the predicted adversarial target.
- Define a calibrated target through a projection:
where is an upper bound on the current estimation error, and controls the aggressiveness of exploiting the forecast.
CoRT then updates according to:
By choosing , CoRT reduces to robust BEST; for large , it increasingly leverages the black-box predictions.
Consistency–Robustness Tradeoff: Theoretical analysis shows that CoRT strictly outperforms BEST when predictions are accurate (“consistency”) and gracefully degrades to BEST’s robust guarantee for inaccurate or adversarial predictions (“robustness”):
where is increasing from 0 to 1 as .
4. Application: Workload Scheduling under Joint SOOTT Objectives
The SOOTT framework is instantiated in a case paper of dynamic resource allocation for mixed elastic/inelastic workloads modeled after real-world AI cluster operations. Here:
- The tracking cost keeps long-term allocated resources aligned with SLA requirements ().
- The adversarial cost penalizes misallocation relative to actual fluctuating inelastic demand ().
- The switching cost limits frequent, abrupt reallocations.
Experimentation with Google Cluster Data demonstrates that:
- Both BEST and CoRT can maintain trajectory tracking (via a windowed average) while avoiding excessive adaptation (switching).
- CoRT exploits ML-based demand forecasts for inelastic workloads, leading to improved overall cost when prediction is accurate, yet remains robust to mispredictions due to the calibration/projection regime.
Parameter variation studies elucidate the core tradeoffs:
| Parameter | Effect |
|---|---|
| Up-weighting adversarial cost prioritizes accuracy in responding to inelastic workload, at possible expense of long-term tracking or smoothness | |
| Increases preference for smooth action sequences, deters rapid response to unexpected changes | |
| Sets window for averaging, trading immediate responsiveness for stability in trajectory tracking | |
| Controls confidence in predictive input in the learning-augmented setting; larger trusts forecasts more |
The system’s total cost function—central to SOOTT—remains:
5. Theoretical Performance Guarantees and Implications
SOOTT algorithms are evaluated via worst-case competitive analysis. The primary guarantee is the degradation factor (DF) for BEST and CoRT relative to IGA:
This guarantee is conservative and remains valid even as the problem parameters (e.g., cost weights, window size) vary, provided the cost parameters are well-conditioned (). When predictions are perfect, the learning-augmented CoRT achieves near-optimal consistency, while for arbitrary (even adversarial) predictions, its cost remains within a quantifiable multiple of the robust baseline.
6. Connections with Prior Work and Extensions
SOOTT generalizes and unifies existing online optimization and tracking methodologies:
- It extends smoothed online learning frameworks (Zhang et al., 2021), augmenting windowed tracking with adversarial and switching penalties under quadratic and polyhedral cost structures.
- The combination of provable guarantees (via calibration to a robust baseline) and the ability to exploit predictive ML models aligns with recent theory for learning-augmented online optimization.
- SOOTT’s structure—explicit tracking, adversarial, and smoothing terms—makes it compatible with the design of resource allocation schemes, multi-target tracking controllers, and real-time scheduling frameworks.
Further research directions include incorporating adaptive recalibration of trust in predictive models, extending analysis to partial-information settings, and exploring distributed/decentralized SOOTT variants for multi-agent or multi-region tracking.
7. Summary Table: Comparative Properties of SOOTT, BEST, CoRT, and IGA
| Algorithm | Requires ? | Uses Black-box Prediction | Robustness Guarantee | Consistency (when predictions are accurate) |
|---|---|---|---|---|
| IGA | Yes | No | Ideal benchmark | Attains lowest possible cost |
| BEST | No | No | DF ≤ explicit bound | Matches robust baseline |
| CoRT | No | Yes | DF ≤ BEST × (1 + O(θ²)) | Approaches IGA as θ↑, remains robust |
8. Conclusion
Smoothed Online Optimization for Target Tracking provides a rigorous, versatile foundation for online decision-making under uncertainty, balancing trajectory tracking, adversarial robustness, and control on decision smoothness. The theoretical and empirical results for BEST and CoRT illustrate both the necessity of robust smoothing and the benefit of integrating (projected) predictions. With its ability to manage nonstationary targets, adversarial disturbances, and dynamically evolving constraints, the SOOTT framework significantly broadens the toolkit for constructing high-reliability, low-regret online tracking systems in demanding real-world applications (Zeynali et al., 7 Sep 2025).
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free