Papers
Topics
Authors
Recent
2000 character limit reached

Multi-objective Bayesian Optimization

Updated 16 December 2025
  • Multi-objective Bayesian Optimization is an advanced framework for identifying Pareto-optimal solutions in expensive black-box settings with multiple conflicting objectives.
  • It employs Gaussian process surrogates and acquisition functions like expected hypervolume improvement to balance exploration and exploitation efficiently.
  • MOBO is applied in engineering, machine learning, and automated design, while addressing challenges in high-dimensional scalability and constraint integration.

Multi-objective Bayesian Optimization (MOBO) addresses the problem of efficiently solving black-box optimization problems involving multiple performance criteria that are expensive to evaluate and often conflicting. Unlike single-objective Bayesian optimization, which seeks a single global extremum, MOBO aims to identify the set of Pareto-optimal solutions—those for which no objective can be improved without degrading another—thus capturing the trade-off surface underlying practical, multi-criteria design and decision-making. MOBO frameworks are increasingly applied in scientific computing, engineering design, machine learning pipeline optimization, and automated composition of complex AI systems (Sabbatella, 14 Nov 2025, Ngo et al., 23 Oct 2025, Li et al., 6 Nov 2024, Martín et al., 2021, Irshad et al., 2022).

1. Mathematical Formulation and Pareto Concepts

In MOBO, the optimization domain is a compact subset XRd\mathcal{X} \subset \mathbb{R}^d, and the objectives are modeled as a vector-valued function f(x)=(f1(x),,fM(x))f(x) = (f_1(x), \ldots, f_M(x)), with fif_i typically expensive and observed only indirectly (e.g., via simulation or real-world experiments). The primary goal is to efficiently approximate the Pareto set

P={xX: xX, fi(x)fi(x), i, j:fj(x)<fj(x)},\mathcal{P}_* = \left\{ x \in \mathcal{X} : \nexists~x' \in \mathcal{X},~f_i(x') \leq f_i(x),~\forall i,~\exists j: f_j(x') < f_j(x) \right\},

and its image, the Pareto front. Pareto dominance is the defining partial order: x1x2x_1 \prec x_2 iff fi(x1)fi(x2)f_i(x_1) \leq f_i(x_2) for all ii and strictly less for at least one ii.

In settings with constraints cj(x)τjc_j(x) \leq \tau_j, the feasible Pareto front is defined over X={xX:cj(x)τj j}\mathcal{X}^* = \{x \in \mathcal{X}: c_j(x) \leq \tau_j~\forall j \}, and new dominance relations are defined with respect to feasibility (Li et al., 6 Nov 2024).

2. Surrogate Modeling and Acquisition Construction

The Bayesian surrogate framework places independent Gaussian process (GP) priors on all objectives,

fi(x)GP(mi(x),ki(x,x)),f_i(x) \sim \mathcal{GP}(m_i(x), k_i(x, x')),

with hyperparameters (mean, kernel, noise variance) learned via marginal likelihood maximization on observed data (Sabbatella, 14 Nov 2025, Roussel et al., 2020, Martín et al., 2021). Constraints, hidden failures, or other black-box properties may be handled by surrogates on additional constraint or classifier outputs (Li et al., 6 Nov 2024, Tran et al., 2020).

The central acquisition design is the expected hypervolume improvement (EHVI):

αEHVI(x)=Ef[max(0,HV(P{f(x)})HV(P))]\alpha_{\text{EHVI}}(x) = \mathbb{E}_f \left[ \max \left(0, HV(P \cup \{f(x)\}) - HV(P) \right) \right]

where PP is the current Pareto set and HVHV denotes Lebesgue measure of the dominated region under a user-supplied reference. EHVI naturally encodes the exploration–exploitation trade-off: high GP mean at unexplored Pareto regions and high uncertainty both receive high scores (Sabbatella, 14 Nov 2025, Roussel et al., 2020, Ngo et al., 23 Oct 2025). UCB variants and scalarization-based schemes, such as random scalarizations sλ(f(x))s_\lambda(f(x)) with λ\lambda sampled over the simplex, provide practical alternatives that scale better to many objectives (Sabbatella, 14 Nov 2025, Paria et al., 2018, Martín et al., 2021).

Acquisition optimization is typically nonconvex and uses multistart gradient methods or evolutionary strategies. For high dimensions, local surrogate models in trust regions are used to manage scaling and mitigate "boundary over-exploration" (Daulton et al., 2021, Uribe-Guerra et al., 19 Sep 2024).

3. Batch, Constrained, and High-Dimensional Extensions

Batch (multi-point) MOBO enables parallel evaluation. Batch EHVI (q-EHVI), batch entropy search, and Kriging Believer strategies have been developed for this setting (Ngo et al., 23 Oct 2025, Wada et al., 2019, Ahmadianshalchi et al., 13 Jun 2024). Determinantal point processes (DPPs) serve to assemble batches that are diverse in input or output space, specifically promoting Pareto front diversity in each iteration (Ahmadianshalchi et al., 13 Jun 2024).

Constrained MOBO methods, for multiple unknown or expensive black-box constraints, include CMOBO, which estimates high-probability confidence intervals for constraints and restricts the search to an optimistic feasible region. Acquisition within this region uses random scalarizations of the objectives constructed to maximize hypervolume. This approach admits finite-sample bounds on both hypervolume regret and cumulative constraint violation (Li et al., 6 Nov 2024).

High-dimensional MOBO poses additional statistical and computational challenges: surrogate model scaling, curse of dimensionality in exploration, and inefficiency of global surrogates. Regionalized or trust region-based frameworks (e.g., MORBO) maintain multiple local GPs and coordinate their exploration via hypervolume contribution maximization, providing improved Pareto coverage, diversity, and practical scaling to hundreds of input features (Daulton et al., 2021, Uribe-Guerra et al., 19 Sep 2024).

4. Algorithmic Frameworks and Notable Variants

The canonical MOBO workflow uses the following template (Sabbatella, 14 Nov 2025, Ngo et al., 23 Oct 2025, Roussel et al., 2020, Martín et al., 2021):

  1. Initialize with a small sample (e.g., Latin Hypercube).
  2. Fit independent GPs to each objective (and to constraint/feasibility indicators as required).
  3. At each iteration:
    • Construct the current Pareto front from all evaluated solutions.
    • Build the acquisition (e.g., EHVI, UCB-variant, random-scalarization, batch entropy).
    • Maximize the acquisition over the design domain (global or local, constrained as necessary) to select one or more candidates.
    • Evaluate the black-box objectives (and constraints) at selected points; update the dataset.
    • Refit GPs as needed.
  4. After budget exhaustion, return the non-dominated set as the estimated Pareto front.

Reducer-based approaches (such as MaO-BO) incorporate automatic objective reduction—removing objectives with similar GP posteriors—improving efficiency in many-objective settings without degrading empirical Pareto fronts (Martín et al., 2021).

Game-theoretic MOBO frameworks allow targeting of specific equilibria, such as Nash, Kalai–Smorodinsky, or Nash–Kalai–Smorodinsky solutions, as alternatives to the global Pareto front, with relevant acquisition rules (UCB-regret, stepwise uncertainty reduction) and theoretical justification (Binois et al., 2021).

Preference-constrained or region-targeted MOBO enables flexible or interactive targeting of front sub-regions either via prior over scalarizations (Paria et al., 2018, Abdolshah et al., 2019, Ozaki et al., 2023) or by fitting a DM's latent utility using pairwise or ordinal queries (Astudillo et al., 20 Jun 2024, Ozaki et al., 2023).

Information-theoretic MOBO variants optimize acquisition functions that maximize the expected information gain about the optimal Pareto region, either via entropy or joint entropy over the Pareto set (PFES, JES) (Suzuki et al., 2019, Tu et al., 2022).

5. Practical Applications and Performance Characterization

MOBO has been demonstrated in domains such as:

Quantitative metrics for MOBO algorithm performance include hypervolume (HV), Diversity Indicator (DIR), IGD/IGD+, spread within objective space, batch efficiency, empirical constraint violation, and regret relative to the true Pareto set or selected equilibria. Recent frameworks consistently outperform evolutionary multi-objective optimization (EMO) baselines in sample efficiency and front diversity, especially for expensive black-box problems with limited evaluations (Ngo et al., 23 Oct 2025, Uribe-Guerra et al., 19 Sep 2024, Ahmadianshalchi et al., 13 Jun 2024, Daulton et al., 2021).

6. Limitations, Open Challenges, and Theoretical Guarantees

The principal challenges for MOBO include:

  • The combinatorial and computational scaling of HV-based acquisitions: exact computation is tractable for M3M\leq3, while for M>3M>3, approximations or submodular surrogates are needed.
  • Surrogate model scalability: managing the cubic scaling in candidate points and multiple objectives is nontrivial for high-dimensional or many-objective problems (Daulton et al., 2021).
  • Expressiveness in constraints and preferences: integration of nontrivial feasibility models, latency or risk-aware objectives, and preference elicitation remains active research (Li et al., 6 Nov 2024, Ozaki et al., 2023, Astudillo et al., 20 Jun 2024).
  • Regret analysis: While sublinear hypervolume regret bounds have been established for EHVI, random-scalarization, and random hypervolume scalarizations in both unconstrained and constrained settings, many recent informational or multi-fidelity MOBO variants lack tight theoretical convergence rates (Paria et al., 2018, Li et al., 6 Nov 2024).

Recent advances include theoretical guarantees on cumulative hypervolume regret and constraint violations under optimistic constraint estimation and random-scalarization acquisitions (Li et al., 6 Nov 2024, Paria et al., 2018), as well as asymptotic consistency for preferential-dueling MOBO with scalarized Thompson sampling (Astudillo et al., 20 Jun 2024). However, extending finite-time regret guarantees to batch, robust, or preference-interactive settings remains open.

7. Future Directions and Prospective Extensions

Notable directions for MOBO research and deployment include:

  • Multi-fidelity and multi-source optimization: leveraging fast proxies, partial evaluations, and hierarchical surrogate modeling to accelerate convergence and reduce expensive true function queries (Irshad et al., 2022, Sabbatella, 14 Nov 2025).
  • Advanced constraint integration: robust handling of black-box, uncertain, or probabilistic constraints, extensions to multi-modal or mixed-variable domains, and non-i.i.d. noise models (Li et al., 6 Nov 2024, Daulton et al., 2022).
  • Automatic, active, and interactive preference learning: engaging decision makers via efficient pairwise, improvement, or dueling feedback; active query selection to reduce interaction cost while focusing on the most relevant trade-offs (Astudillo et al., 20 Jun 2024, Ozaki et al., 2023).
  • Regionalization and trust-region architectures for scaling to ultra-high-dimensional design problems, with dynamic resource allocation and adaptive exploration–exploitation balancing (Daulton et al., 2021, Uribe-Guerra et al., 19 Sep 2024).
  • Integration with game-theoretic solutions for applications where fairness or multi-agent equilibrium is the desired optimization target (Binois et al., 2021).

A plausible implication is that the evolution of MOBO methods is increasingly dictated by the demands of scale, data fit, real-world constraints, and human-in-the-loop utility maximization, as evidenced by the expanding repertoire of regularized surrogates, information-theoretic acquisitions, preference-based learning, and robust optimization architectures (Sabbatella, 14 Nov 2025, Li et al., 6 Nov 2024, Ngo et al., 23 Oct 2025, Uribe-Guerra et al., 19 Sep 2024, Astudillo et al., 20 Jun 2024, Ozaki et al., 2023).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)

Whiteboard

Follow Topic

Get notified by email when new papers are published related to Multi-objective Bayesian Optimization.