Papers
Topics
Authors
Recent
Search
2000 character limit reached

Multi-Objective Bayesian Optimization

Updated 20 January 2026
  • Multi-Objective Bayesian Optimization is a probabilistic framework that integrates independent Gaussian process surrogates and acquisition functions to balance exploration and exploitation in expensive multi-objective problems.
  • It employs techniques like hypervolume improvement, random scalarizations, and information-theoretic methods to efficiently explore and approximate Pareto-optimal sets with provable theoretical guarantees.
  • MOBO finds applications in engineering, drug discovery, and system design while addressing challenges such as constraints handling, preference elicitation, and scalability in high-dimensional spaces.

Multi-objective Bayesian optimization (MOBO) is a probabilistic framework for finding the Pareto-optimal set of solutions to vector-valued black-box optimization problems, where function evaluations are expensive. MOBO combines surrogate modeling (typically independent Gaussian processes for each objective), acquisition functions tailored to trade off exploration and exploitation under multiple objectives, and sequential experimental design to efficiently approximate the Pareto front or a specific preferred solution. The field encompasses constrained settings, input uncertainty, many-objective regimes, preference elicitation, batch evaluations, and high-dimensional search spaces, and is motivated by applications in scientific experiment design, engineering, drug discovery, and system optimization.

1. Formulation and Pareto Optimality in MOBO

The central task in MOBO is to optimize mm black-box objectives F(x)=[f1(x),,fm(x)]F(x) = [f_1(x),\dots,f_m(x)] over a compact or finite domain XRd\mathcal X\subset\mathbb R^d, often subject to cc black-box constraints G(x)=[g1(x),...,gc(x)], gj(x)0G(x) = [g_1(x),...,g_c(x)],\ g_j(x)\ge0. The feasible region is XF={xX:gj(x)0 j}\mathcal X_F = \{x \in \mathcal X: g_j(x)\ge0\ \forall j\}. Pareto optimality under constraints is defined such that xx Pareto-dominates xx' if fi(x)fi(x)f_i(x) \ge f_i(x') for all ii and strictly >> for at least one ii, with both x,xx,x' feasible. The goal is to identify the (possibly constrained) Pareto set XF\mathcal X^*_F and its image P={F(x):xXF}\mathcal P = \{ F(x) : x \in \mathcal X^*_F \} (Li et al., 2024).

Performance is frequently measured by the hypervolume (HV) indicator, HVz(P)\mathcal{HV}_z(\mathcal P), which quantifies the Lebesgue measure of the dominated region with respect to a reference point zz; simple regret, generational distance (GD), and inverted GD (IGD) are also common.

2. Surrogate Modeling and Acquisition Function Design

MOBO employs independent (or, in certain cases, multi-output) Gaussian process (GP) priors for each fif_i and, when present, each gjg_j. After t1t-1 evaluations, the GP posterior yields predictive means μh,t1(x)\mu_{h,t-1}(x) and variances σh,t12(x)\sigma^2_{h,t-1}(x) for h{fi,gj}h \in \{ f_i, g_j \}. Acquisition functions—quantifying the expected utility or information gain from evaluating xx—are critical for sequential design. Categories include:

Constraint handling leverages optimistic GP-UCB surrogates; in CMOBO, constraint acquisition is restricted to inputs with constraint-UCBs above threshold ("optimistic constraint estimation") (Li et al., 2024). Robust MOBO frameworks under input noise deploy Bayes risk surrogates (Qing et al., 2022), or optimize multivariate value-at-risk (MVaR) via random scalarizations (Daulton et al., 2022).

3. Theoretical Properties and Sample Efficiency Guarantees

Rigorous convergence and regret analyses are available for several classes of MOBO algorithms, most notably for settings using optimistic constraint-UCB with random-hypervolume scalarization. Under standard GP-UCB assumptions (RKHS boundedness, Gaussian noise, Lipschitzity of scalarizations), the cumulative hypervolume regret after TT iterations satisfies

RTO(m2γTTlnT)\mathcal R_T \le O(m^2 \sqrt{\gamma_T T \ln T})

where γT\gamma_T is the maximum information gain permitted by the GPs (Li et al., 2024). For constraints, cumulative violation is

Vj,TO(γTTlnT)\mathcal V_{j,T} \le O(\sqrt{\gamma_T T \ln T})

and the best combined regret is O((c+m2)γTlnT/T)O((c+m^2)\sqrt{\gamma_T \ln T / T}).

Infeasibility detection is also covered: with high probability, the algorithm does not falsely declare infeasibility when feasible solutions exist, and reliably detects infeasibility (within quantifiable iteration bounds) when the feasible set is empty.

When random scalarizations are used (linear, Chebyshev, or mixed), sublinear Bayes (simple) regret is provable: E[RC(T)]=O(LKTdγTlnT)E[R_C(T)] = O(L \cdot K \cdot \sqrt{T d \gamma_T \ln T}) where LL is the Lipschitz constant of the scalarizations and KK is the number of objectives (Paria et al., 2018).

MOBO under robust objectives (Bayes risk, MVaR) can achieve strong performance guarantees by virtue of bijective correspondences between Chebyshev scalarizations and robust Pareto set points (Daulton et al., 2022).

4. Algorithmic Implementations and Computational Considerations

A typical MOBO algorithmic loop comprises:

  1. GP fitting: Train independent (or multi-output) GPs for all objectives and constraints using the data collected so far.
  2. Acquisition optimization: For either a batch or a single candidate, solve for xx maximizing the acquisition within the current possibly-feasible set. Constrained regions are estimated optimistically using the UCBs of constraint surrogates (Li et al., 2024).
  3. Constraint handling: Acquisition maximization is constrained to xx satisfying UCBgj(x)0\mathrm{UCB}_{g_j}(x)\geq 0 for all jj (or an appropriate confidence set).
  4. Hypervolume scalarization: At each step, sample a random scalarization direction θ\theta and use the corresponding acquisition sθs_\theta on the GP-UCB surrogates, maximizing the scalarized UCB subject to constraint-UCBs (Li et al., 2024).
  5. Candidate evaluation: Evaluate F(x),G(x)F(x),G(x) at the selected xx and update the GP posteriors.
  6. Global infeasibility check: If no xx satisfies the optimism constraints, declare the problem globally infeasible.

The computational bottleneck is dominated by GP inference (O(nm2nm^2)), acquisition maximization (often high due to the need for global optimization of non-convex surrogates), and, for hypervolume- or entropy-based acquisitions, combinatorial region decompositions in high mm (Tu et al., 2022, Suzuki et al., 2019). Batched MOBO techniques use greedy or Kriging-believer selection, and batch-diversity is maximized via DPPs (Ahmadianshalchi et al., 2024), or orthogonal search directions (Ngo et al., 23 Oct 2025).

Scalability in high-dimensional spaces is achieved by regionalized surrogates, local trust-regions, and data-sharing, as in MORBO (Daulton et al., 2021).

5. Extensions: Preference Learning, Robustness, and Many-Objective Regimes

Preference-Aware BO: Active preference learning (pairwise comparisons and improvement requests) with Bayesian preference models over the decision-maker's utility vector ww enables query-efficient identification of the most preferred Pareto-optimal solution. Acquisition incorporates both GP and utility-parameter uncertainty, and active learning selects informative preference queries via mutual information (BALD) (Ozaki et al., 2023).

Preference-Order Constraints: MOBO-PC formulates polyhedral cones in the space of objective weights to encode stability preferences (e.g., "objective A is more important than B"). The acquisition function uses a weighted expected hypervolume improvement, where the weight is the probability (under a GP gradient surrogate) that a solution satisfies the preference constraints (Abdolshah et al., 2019).

Robustness to Input Uncertainty: Robust MOBO frameworks quantify performance under uncertain inputs via Bayes risk (expectation over input perturbation) (Qing et al., 2022) or by optimizing the multivariate value-at-risk (MVaR), approximated by random Chebyshev scalarization and appropriate risk quantiles (Daulton et al., 2022).

Many-Objective BO: For large mm, redundancy in objectives is detected via GP-predictive similarity metrics, allowing dynamic pruning of objectives and significant computational/resource savings without loss of Pareto diversity (Martín et al., 2021).

6. Empirical Performance, Applications, and Benchmarks

CMOBO, with optimistic constraint estimation and random hypervolume scalarization, achieves or surpasses state-of-the-art performance in both synthetic (toy, Branin–Currin, C2-DTLZ2) and real-world (penicillin production, disc brake design, Caco-2++ drug discovery, ESOL+ solubility) settings, as measured by hypervolume, cumulative violation, and constraint regret (Li et al., 2024). Results demonstrate:

Method Hypervolume ↑ Violation ↓ Constraint Regret ↓
CMOBO best low best
qNEHVI good high worse
qParEGO moderate med. med.
MESMOC slow ≈0 med.
Random worst random worst

CMOBO achieves the lowest cumulative violation, with MESMOC being overly conservative and slower in hypervolume improvement. CMOBO's performance is robust on high-dimensional chemical inputs (2,100–2,200++D).

Other advanced MOBO methods such as Joint Entropy Search (JES) (Tu et al., 2022), PFES (Suzuki et al., 2019), and scalable regional trust-region BO (MORBO) (Daulton et al., 2021) demonstrate comparable or superior performance on standard and large-scale benchmarks, substantially surpassing heuristic baselines.

7. Future Research Directions and Open Challenges

Challenges for MOBO include rigorous scaling to extremely high-dimensional spaces (jointly in input and objective), robust optimization under complex input uncertainty models, theoretical convergence in the presence of cycles and complex networked objectives (Kudva et al., 19 Feb 2025), practical interactive preference learning at scale, and the integration of multiple constraints and domain constraints with provable guarantees. Automated selection of scalarizations, acquisition functions, and prioritization among objectives remains an open issue. Ongoing developments in batch/diversity-aware optimization, robust acquisition under risk, and automated algorithm configuration continue to advance the state of the art in MOBO research.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Multi-Objective Bayesian Optimization.