SOTAVerified

A new integral loss function for Bayesian optimization

2014-08-20Unverified0· sign in to hype

Emmanuel Vazquez, Julien Bect

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We consider the problem of maximizing a real-valued continuous function f using a Bayesian approach. Since the early work of Jonas Mockus and Antanas Zilinskas in the 70's, the problem of optimization is usually formulated by considering the loss function f - M_n (where M_n denotes the best function value observed after n evaluations of f). This loss function puts emphasis on the value of the maximum, at the expense of the location of the maximizer. In the special case of a one-step Bayes-optimal strategy, it leads to the classical Expected Improvement (EI) sampling criterion. This is a special case of a Stepwise Uncertainty Reduction (SUR) strategy, where the risk associated to a certain uncertainty measure (here, the expected loss) on the quantity of interest is minimized at each step of the algorithm. In this article, assuming that f is defined over a measure space (X, ), we propose to consider instead the integral loss function _X (f - M_n)_+\, d, and we show that this leads, in the case of a Gaussian process prior, to a new numerically tractable sampling criterion that we call EI^2 (for Expected Integrated Expected Improvement). A numerical experiment illustrates that a SUR strategy based on this new sampling criterion reduces the error on both the value and the location of the maximizer faster than the EI-based strategy.

Tasks

Reproductions