Papers
Topics
Authors
Recent
Search
2000 character limit reached

Adaptive Meta Black-box Optimization (ABOM)

Updated 5 February 2026
  • ABOM is a meta-optimization framework that integrates offline meta-learning, surrogate modeling via an Attentive Neural Process, and Bayesian optimization for rapid adaptation to unseen tasks.
  • It constructs a diverse meta-dataset through parallel simulation and uses well-calibrated uncertainty estimates to significantly reduce the number of expensive online evaluations.
  • Empirical evaluations in urban traffic management demonstrate ABOM's ability to lower waiting vehicles and improve throughput with fewer than 100 trials.

The Adaptive Meta Black-box Optimization Model (ABOM) formalizes a class of meta-optimization frameworks that leverage offline meta-learning and modern surrogate modeling to enable rapid, data-efficient adaptation to unseen black-box optimization tasks. The paradigm is motivated by real-world applications such as urban traffic network design, where conventional optimization or heuristic controllers fail to generalize due to heterogeneity and limited sample budgets. ABOM, as introduced in traffic light management (Yun et al., 2024), combines an offline meta-dataset of task–design–performance pairs, a meta-learned Attentive Neural Process (ANP) surrogate, and Bayesian optimization to deliver sample-efficient, uncertainty-aware optimization of system-level designs such as traffic phase configurations and timing.

1. Formulation of the Meta-Black-box Optimization Problem

ABOM considers a family of black-box optimization tasks defined by an unknown distribution ρ. Each task (e.g., a traffic pattern) p induces a black-box objective fp:RdRf_p: \mathbb{R}^d \to \mathbb{R} that evaluates high-level design variables xx (e.g., intersection phase combinations, green-time allocations). The core challenge is to efficiently optimize x=argmaxxfp(x)x^* = \arg\max_x f_p(x) for a new, unseen pattern punseenp^{unseen}, using as few expensive function evaluations as possible.

ABOM operates in a meta-learning regime: it is provided with an offline meta-dataset D={Dn}n=1ND = \{ D^n \}_{n=1}^N, where each Dn={(xmn,ymn)}m=1MD^n=\{(x^n_m, y^n_m)\}_{m=1}^M comprises design–performance pairs (x,y=fpn(x)+ϵ)(x, y = f_{p^n}(x)+\epsilon) collected under NN historical tasks (traffic scenarios). This meta-dataset is used to inform and “warm-start” the search for xx^* in a new task punseenp^{unseen}, given a limited budget KK of online simulations.

Key components:

  • Variable: global traffic-light design vector xRdx \in \mathbb{R}^{d}, with d=IJd=I \cdot J for II intersections and JJ choices each.
  • Objective: fp(x)f_p(x), noisy, nonconvex, and non-differentiable.
  • Data: DD aggregates prior realizations of (x,fpn(x))(x, f_{p^n}(x)) for n=1,,Nn=1,\ldots,N.
  • Goal: Use DD to minimize the number of online evaluations needed to solve argmaxxfpunseen(x)\arg\max_x f_{p^{unseen}}(x).

2. Construction and Role of the Offline Meta-Dataset

Data collection is performed entirely offline in parallelized simulation. For each reference pattern pnp^n, MM candidate designs xmnx^n_m are sampled uniformly from the design space (phase-combination one-hot logits, or unnormalized green-time splits). Their simulated performance ymn=fpn(xmn)+ϵy^n_m = f_{p^n}(x^n_m) + \epsilon yields DnD^n without sequential dependency, ensuring the dataset is unbiased and exploits parallel compute.

This process is repeated independently for each pnp^n (N=120N=120 in large-scale experiments; M=200M = 200), yielding a meta-dataset D={Dn}n=1ND = \{ D^n \}_{n=1}^N that covers a diverse spectrum of traffic conditions and associated interventions. The data supports learning a prior over the mapping from designs to outcomes, enabling generalization to unseen traffic scenarios in a principled few-shot adaptation framework.

A held-out validation split ($5:1$ train/val over NN) is used for hyperparameter tuning of the meta-surrogate.

3. Attentive Neural Process Surrogate: Architecture and Training

3.1 Latent Neural Process Structure

The ABOM surrogate is an Attentive Neural Process (ANP), modeling the conditional stochastic process fρf \sim \rho:

  • For any context set C={(xc,yc)}C = \{(x_c, y_c)\} and any target set T={(xt,yt)}T = \{(x_t, y_t)\}:
    • ANP introduces a global latent vector zz with prior qϕ(zC)q_\phi(z|C) and defines:

    p(zC)tTp(ytxt,z)p(z|C)\prod_{t \in T} p(y_t|x_t, z)

  • Encoder: Eϕ(xc,yc)rcE_\phi(x_c, y_c) \to r_c for each context; updated via self-attention to R=Attnself({rc})R = Attn_{self}(\{r_c\}) and aggregated rC=mean(R)r_C = \text{mean}(R).

  • Recognition model: qϕ(zC)=N(μz(rC),σz2(rC))q_\phi(z|C) = \mathcal{N}(\mu_z(r_C), \sigma_z^2(r_C)).

  • Decoder: pθ(yx,z)=N(μθ(x,z,ax),σθ2(x,z,ax))p_\theta(y|x, z) = \mathcal{N}(\mu_\theta(x, z, a_x), \sigma_\theta^2(x, z, a_x)), where axa_x is a cross-attention over RR conditioned on query xx.

3.2 Training Objective

The surrogate is meta-trained across all NN tasks to maximize the evidence lower bound (ELBO):

L(ϕ,θ)Eqϕ(zT)[tTlogpθ(ytxt,z)]KL[qϕ(zT)qϕ(zC)]\mathcal{L}(\phi,\theta) \geq \mathbb{E}_{q_\phi(z|T)} \left[ \sum_{t \in T} \log p_\theta(y_t|x_t, z) \right] - \mathrm{KL}[q_\phi(z|T) \| q_\phi(z|C)]

This process alternates between random context/target splits of each DnD^n, encouraging the model to learn transferable structure for rapid adaptation. The ANP models predictive mean and variance, naturally encoding epistemic uncertainty critical for Bayesian optimization.

3.3 Inference

During online adaptation on punseenp^{unseen}, the ANP receives DonD_{on} (the current online trials) as context and returns:

  • Predictive mean: μ(xDon)=Eq(zDon)[μθ(x,z,ax)]\mu(x|D_{on}) = \mathbb{E}_{q(z|D_{on})}[\mu_\theta(x, z, a_x)]

  • Predictive variance: σ2(xDon)=Varq(zDon)[μθ(x,z,ax)]+Eq(zDon)[σθ2(x,z,ax)]\sigma^2(x|D_{on}) = \operatorname{Var}_{q(z|D_{on})}[\mu_\theta(x, z, a_x)] + \mathbb{E}_{q(z|D_{on})}[\sigma^2_\theta(x, z, a_x)] These are estimated via Monte Carlo or moment matching.

4. Bayesian Optimization with the ANP Surrogate

ABOM applies Bayesian optimization (BO) to maximize fpunseen(x)f_{p^{unseen}}(x) using the ANP as a probabilistic surrogate. Standard acquisition functions are employed:

aUCB(x)=μ(xDon)+βσ(xDon),β=2.0a_{UCB}(x) = \mu(x|D_{on}) + \beta \sigma(x|D_{on}), \quad \beta = 2.0

  • Expected Improvement (EI):

Z=μ(x)f+σ(x),aEI(x)=(μ(x)f+)Φ(Z)+σ(x)ϕ(Z)Z = \frac{\mu(x) - f^+}{\sigma(x)}, \quad a_{EI}(x) = (\mu(x) - f^+) \Phi(Z) + \sigma(x) \phi(Z)

  • Probability of Improvement (PI):

aPI(x)=Φ(μ(x)f+σ(x))a_{PI}(x) = \Phi\left(\frac{\mu(x) - f^+}{\sigma(x)}\right)

Constraints are enforced:

  • For phase time allocation: simplex constraints on each xix_i: jsoftmax(xij)=1\sum_j \mathrm{softmax}(x_{ij}) = 1 and softmax(xij)(cJgmin)+gmin0\mathrm{softmax}(x_{ij}) \cdot (c - J g_{min}) + g_{min} \geq 0.

  • For phase combinations: argmax over JJ softmax probabilities ensures valid discrete selection.

Acquisition maximization over feasible xx is solved with L-BFGS-B on the continuous logits.

Optimization Loop

At each online trial:

  1. Compute surrogate mean/variance given current DonD_{on}.

  2. Maximize the acquisition function a(x)a(x) under the constraints to propose next xx.

  3. Evaluate fpunseen(x)f_{p^{unseen}}(x) via simulation.

  4. Augment online context with new (x,y)(x, y). The best xx found is returned.

5. Algorithmic Workflow and Pseudocode

Phase 1: Offline Data Collection

  • For n=1n=1 to NN: sample MM designs, evaluate under fpnf_{p^n}, store in DnD^n.

Phase 2: Meta-training ANP

  • Until convergence: randomly select DnD^n, split into context/target, update ANP via stochastic gradient ascent on the ELBO.

Phase 3: Online Adaptation on New Task

  • Initialize Don=D_{on} = \emptyset.

  • For k=1,,Kk = 1,\ldots,K:

    • Compute μ(x),σ(x)\mu(x),\sigma(x) given ANP and DonD_{on}.
    • Maximize acquisition a(x)a(x) (UCB, EI, or PI) under constraints.
    • Query fpunseen(x)f_{p^{unseen}}(x), record (x,y)(x, y).
  • Return xx with best observed fpunseen(x)f_{p^{unseen}}(x).

6. Empirical Evaluation and Impact

ABOM was instantiated on both synthetic 2×22 \times 2 to 5×55 \times 5 grids and real networks (Hangzhou 4×44 \times 4, Manhattan 16×316 \times 3 and 28×728 \times 7). Meta-dataset: N=120N=120, M=200M=200. Baselines included basic BBO (GA, PSO, CMA-ES), meta BBO (LGA, LES, RGPE, ABLR, FSBO), and RL 1-step MDP controllers (e.g., DQN, PPO).

Key results:

  • On Hangzhou 4×44 \times 4 (phase combination), ABOM: $292.7$ vs. best baseline $294.8$ average waiting vehicles.
  • On Manhattan 28×728 \times 7 (time allocation), ABOM: $3805.4$ vs. $3890.4$ next best.
  • ABOM reduced waiting vehicles by 210%2-10\% across all networks, converged in <100<100 trials.
  • Real deployment: 26 intersections, 4.80%4.80\% vehicle throughput improvement over baseline when run hourly for a week.
  • Ablations: robust to halved meta-dataset, all three acquisition criteria give similar improvements, ANP preferred over transformer NPs for the available data scale.

7. Interpretation and Significance

ABOM demonstrates that offline meta-learning of surrogate models (ANP) over diverse design–performance datasets, combined with probabilistic BO, enables high sample-efficiency black-box optimization in complex, real-world system design settings. The approach achieves rapid adaptation beyond traditional BBO and RL variants, robustly outperforms a range of meta and naïve baselines on metrics relevant to practice (e.g., waiting vehicles, throughput), and is deployable at urban scale without domain-specific tuning (Yun et al., 2024). The framework is general and could be extended to other black-box systems where acquiring task-generalizable learned surrogates with well-calibrated uncertainty is feasible.

Empirical findings confirm the following:

  • Parallel offline data generation (random design sampling) is highly effective for meta-surrogate training and allows for maximal use of simulation infrastructure.
  • Well-calibrated epistemic uncertainty from the ANP is critical for BO sample efficiency, irrespective of acquisition function.
  • Robustness to meta-dataset size and acquisition choice confirms that meta-BBO with few-shot adaptation is a practical architecture for scalable, adaptive design in high-stake black-box domains.

Principal Reference:

"An Offline Meta Black-box Optimization Framework for Adaptive Design of Urban Traffic Light Management Systems" (Yun et al., 2024)

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Adaptive Meta Black-box Optimization Model (ABOM).