Adaptive and Aggressive Rejection (AAR)
- AAR is a dynamic rejection framework that filters anomalous or adversarial contributions using robust statistical thresholds and a Gaussian mixture model for soft rejection.
- It integrates a warm-up phase with hard rejection and a main phase with ternary weighting to optimize data retention and improve performance metrics like AUROC.
- In adaptive control, AAR employs disturbance observers and finite-time controllers to aggressively cancel perturbations, ensuring rapid convergence under uncertainty.
Adaptive and Aggressive Rejection (AAR) encompasses a family of algorithmic mechanisms for robustly filtering out undesirable or adversarial contributions during inference or learning. In anomaly detection, AAR refers to a dynamic, data-driven rejection framework that adaptively identifies and excludes contaminated samples by jointly leveraging robust statistical thresholds and probabilistic modeling. In nonlinear control, AAR describes the coordinated use of adaptive, experience-accelerated disturbance estimators and finite-time controllers to aggressively cancel exogenous perturbations. Across both contexts, AAR is characterized by its principled, multi-phase rejection logic and its capacity to dynamically optimize the trade-off between retention and exclusion under uncertainty.
1. Mathematical Foundations of AAR for Anomaly Detection
AAR for anomaly detection operates on a contaminated dataset , using anomaly scores . For reconstruction-based models, %%%%2%%%%, where is typically an autoencoder. The framework dynamically rejects anomalies in each mini-batch via a tiered thresholding procedure:
- Modified z–score (hard rejection): For batch size , compute
Samples with (i.e., ) are hard rejected with threshold
- Gaussian Mixture Model (GMM) intersection (soft rejection): Fit a two-component GMM,
The intersection threshold solves
Explicitly,
with
- – (stability guard): Compute
with the mean/std of the “normal” GMM component, .
The final soft rejection threshold is .
2. Integrated Hard and Soft Rejection Strategies
AAR integrates these thresholds into a phased rejection weighting scheme:
- Warm-up (first epochs): Only the hard cutoff is active; for , otherwise.
- Main phase (): Use weights
with (typically $0.1$).
This approach transforms sample selection from binary (keep/discard) into a ternary regime , allowing ambiguous samples to influence training with attenuated impact. This aggressive rejection—removing more than the nominal contamination—empirically yields heightened robustness and improved AUROC, particularly when normal and anomaly score distributions overlap (Lee et al., 26 Nov 2025).
3. AAR Algorithm and Computational Complexity
The AAR training protocol proceeds as follows for each mini-batch and epoch up to :
- Compute anomaly scores .
- Determine for all epochs.
- If :
- Fit GMM and derive , , .
- Assign weights according to the current phase.
- Compute the weighted loss,
and update the model.
Computationally, for mini-batch size and feature dimensionality , the per-step cost is dominated by the forward/backward pass ; thresholding and EM for GMM fitting add overhead, rendering AAR scalable for large and .
4. AAR in Adaptive Control: Disturbance Rejection
In adaptive nonlinear control, AAR is exemplified by architectures that combine online disturbance identification with aggressive, finite-time error suppression (Li et al., 2020). Consider a nonlinear plant,
subject to exosystem-generated disturbance . The core components are:
- Adaptive disturbance observer: State-derivative-free estimation using a filtered regressor, adaptive update of via Lyapunov-stable adaptation,
where experience replay () accelerates convergence.
- Aggressive (finite-time) controller: Integral-type terminal sliding mode with adaptive gain,
enforcing in finite time provided certain rank/richness conditions are met.
The “adaptive” aspect derives from online parameter learning, while “aggressive rejection” is realized through high-bandwidth feed-forward cancellation and non-asymptotic convergence guarantees.
5. Empirical Evaluation and Performance
In anomaly detection benchmarks (Lee et al., 26 Nov 2025), AAR demonstrates:
- On MNIST/Fashion-MNIST with up to synthetic contamination, AAR achieves average AUROC increases of $0.006$ (MNIST) and $0.016$ (F-MNIST) over the prior best latent outlier exposure (LOE).
- On $30$ UCI-type tabular datasets contaminated at , AAR lifts average AUROC by across AE/MemAE/DSVDD backbones relative to robust statistics (MZ).
- Overall, AAR’s average AUROC gain over all prior methods is .
Ablations confirm that slightly over-estimating contamination (by ) enhances robustness; increasing in the – cutoff improves stability with negligible loss; soft rejection with optimizes the bias-variance trade-off.
In adaptive control (Li et al., 2020), experience replay reduces disturbance estimation time from s to s and ensures finite-time tracking in s (in nonlinear benchmarks), contrasting with much slower convergence in experience-free observers.
6. Practical Tuning, Limitations, and Outlook
Tuning recommendations for anomaly detection include (warm-up epochs), (stability guard), and (soft rejection weight). For adaptive control, filter and adaptation gains are selected to balance estimation speed and sensitivity to noise, while the replay window is optimized against memory and numerical stability.
Notable limitations:
- The univariate GMM used in AAR assumes a bimodal, near-Gaussian score distribution, which can be violated in highly skewed or multi-modal cases.
- Hyperparameters still require domain-specific tuning.
Open research directions involve meta-learning for automatic parameter adaptation, integrating limited anomaly labels (semi-supervised AAR), extending to high-dimensional or non-Gaussian score spaces, and adapting AAR for real-time data streams in cyber-physical systems and IoT scenarios (Lee et al., 26 Nov 2025, Li et al., 2020).