Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RTDK-BO: High Dimensional Bayesian Optimization with Reinforced Transformer Deep kernels (2310.03912v5)

Published 5 Oct 2023 in cs.LG and cs.AI

Abstract: Bayesian Optimization (BO), guided by Gaussian process (GP) surrogates, has proven to be an invaluable technique for efficient, high-dimensional, black-box optimization, a critical problem inherent to many applications such as industrial design and scientific computing. Recent contributions have introduced reinforcement learning (RL) to improve the optimization performance on both single function optimization and \textit{few-shot} multi-objective optimization. However, even few-shot techniques fail to exploit similarities shared between closely related objectives. In this paper, we combine recent developments in Deep Kernel Learning (DKL) and attention-based Transformer models to improve the modeling powers of GP surrogates with meta-learning. We propose a novel method for improving meta-learning BO surrogates by incorporating attention mechanisms into DKL, empowering the surrogates to adapt to contextual information gathered during the BO process. We combine this Transformer Deep Kernel with a learned acquisition function trained with continuous Soft Actor-Critic Reinforcement Learning to aid in exploration. This Reinforced Transformer Deep Kernel (RTDK-BO) approach yields state-of-the-art results in continuous high-dimensional optimization problems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Alexander Shmakov (19 papers)
  2. Avisek Naug (15 papers)
  3. Vineet Gundecha (16 papers)
  4. Sahand Ghorbanpour (13 papers)
  5. Ricardo Luna Gutierrez (11 papers)
  6. Ashwin Ramesh Babu (20 papers)
  7. Antonio Guillen (11 papers)
  8. Soumyendu Sarkar (17 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.