Rylan Schaeffer

Logo
Resume
Research
Learning
Blog
Teaching
Jokes
Kernel Papers


Introduction to Bayesian Optimization

Bayesian Optimization (BO) concerns gradient-free, assumption-free optimization. The goal is to solve a general optimization problem with no known structure (e.g. convexity or linearity) to exploit, and where we do not have access to any of the function \(f(\dot)\)’s derivatives.

\[\max_{x \in X} f(x)\]

Many other assumptions are also often applied, including:

All/most approaches to Bayesian Optimization have two components:

  1. A statistical model of the objective function, often called the surrogate function
  2. A method for deciding where to sample next, often called the acquisition function.

There are a variety of different acquisition functions, but almost always, the surrogate function is Gaussian Process regression.

The psuedo-code is then: