Papers
Topics
Authors
Recent
Search
2000 character limit reached

Fast Calculation of the Knowledge Gradient for Optimization of Deterministic Engineering Simulations

Published 16 Aug 2016 in cs.CE, cs.LG, and stat.ML | (1608.04550v1)

Abstract: A novel efficient method for computing the Knowledge-Gradient policy for Continuous Parameters (KGCP) for deterministic optimization is derived. The differences with Expected Improvement (EI), a popular choice for Bayesian optimization of deterministic engineering simulations, are explored. Both policies and the Upper Confidence Bound (UCB) policy are compared on a number of benchmark functions including a problem from structural dynamics. It is empirically shown that KGCP has similar performance as the EI policy for many problems, but has better convergence properties for complex (multi-modal) optimization problems as it emphasizes more on exploration when the model is confident about the shape of optimal regions. In addition, the relationship between Maximum Likelihood Estimation (MLE) and slice sampling for estimation of the hyperparameters of the underlying models, and the complexity of the problem at hand, is studied.

Citations (2)

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Explain it Like I'm 14

Overview

This paper is about a smart way to run computer experiments that take a long time to compute (like detailed engineering simulations) so that we can find the best design using as few runs as possible. The authors focus on a strategy called the Knowledge Gradient for Continuous Parameters (KGCP) and show a fast way to compute it when the simulations are deterministic (they always give the same result for the same input). They compare KGCP with other popular strategies—Expected Improvement (EI) and Upper Confidence Bound (UCB)—and test them on several problems, including a real engineering example.

What questions does the paper try to answer?

Here are the main questions the authors explore:

  • Can we compute the KGCP strategy quickly for simulations that have no noise?
  • How is KGCP different from EI, and when is it better?
  • How does KGCP compare to EI and UCB on easy and hard optimization problems?
  • What’s the best way to set the “tuning knobs” (hyperparameters) of the prediction model: using one best guess (MLE) or by averaging many plausible guesses (slice sampling)?

Methods and approach (in everyday language)

Think of searching for treasure on a huge map, where each dig is expensive. You want a smart plan to decide where to dig next.

  • Surrogate model (Kriging): Instead of digging everywhere, you draw a smooth “prediction map” from the points you’ve already dug. This map estimates the treasure value (the simulation output) at new places and also how uncertain that estimate is. Kriging is a common tool for this: it predicts a value and how confident it is about that prediction.
  • Strategies for choosing the next dig:
    • EI (Expected Improvement): Picks the next spot by estimating “how much better than our current best” we might get if we dig there. It likes spots that look promising or uncertain (because uncertainty means there’s a chance they could be great).
    • UCB (Upper Confidence Bound): A simple rule that balances the predicted value and uncertainty. It tends to exploit promising areas but can get stuck.
    • KGCP (Knowledge Gradient for Continuous Parameters): Chooses the next spot by estimating how much our best decision would improve after learning the result at that spot. It’s more cautious: if the model is already very sure a region is good, KGCP doesn’t waste more digs there—it explores other areas instead.
  • Deterministic setting: In many physics-based simulations, the same input gives the same output every time (no randomness). The authors derive a neat, fast formula to compute KGCP in this case, so it’s about as quick to use as EI.
  • Setting the model’s “tuning knobs” (hyperparameters):
    • MLE (Maximum Likelihood Estimation): Pick the single best set of hyperparameters that makes the model fit the data.
    • Slice sampling: Try many plausible hyperparameter settings and average them, to capture our uncertainty about the model’s settings. This can help when the problem is complex and one “best” setting isn’t reliable.

Main findings and why they matter

Here are the key results from tests on both synthetic functions and a real 10-dimensional engineering problem (a vibrating truss structure):

  • Speed: The new KGCP formula is fast to compute for deterministic problems—about as fast as EI—so it’s practical to use.
  • Performance on easy problems (smooth surfaces like Branin and Hartmann):
    • KGCP and EI perform similarly.
    • UCB can do fine on very simple cases but often isn’t as reliable overall.
  • Performance on hard, bumpy problems (many peaks and valleys, like Eggholder and Schwefel):
    • KGCP tends to find good solutions faster than EI because it explores more when the model is confident in some areas. This reduces the chance of getting stuck in a local (not global) optimum.
    • UCB often gets trapped in local optima on these hard problems.
  • Hyperparameter choices:
    • On smoother, simpler synthetic problems, using MLE (one best set of hyperparameters) was generally as good or better than slice sampling.
    • On the complex 10D truss problem, slice sampling helped, and KGCP with slice sampling performed best overall, continuing to improve when EI had leveled off.

Why this matters: If you’re running expensive simulations, choosing KGCP can help you find the best design in fewer runs, especially when the problem is complicated and has many competing “good” regions.

Implications and potential impact

  • For engineers and scientists optimizing expensive, deterministic simulations, KGCP is now a practical choice because it’s fast to compute.
  • KGCP’s “trust the model, explore elsewhere” behavior can speed up finding the global best, especially on complex, multi-peak problems.
  • How you set the model’s hyperparameters depends on the problem:
    • If the problem is simple and the model’s best settings are clear, MLE is efficient and effective.
    • If the problem is complex and the model’s settings are uncertain, slice sampling can improve results, and pairing it with KGCP is especially strong.
  • Overall, this work suggests using KGCP for challenging engineering optimizations and choosing between MLE or slice sampling based on how complicated and uncertain the modeling is.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.