SOTAVerified

Bayesian Optimisation

Expensive black-box functions are a common problem in many disciplines, including tuning the parameters of machine learning algorithms, robotics, and other engineering design problems. Bayesian Optimisation is a principled and efficient technique for the global optimisation of these functions. The idea behind Bayesian Optimisation is to place a prior distribution over the target function and then update that prior with a set of “true” observations of the target function by expensively evaluating it in order to produce a posterior predictive distribution. The posterior then informs where to make the next observation of the target function through the use of an acquisition function, which balances the exploitation of regions known to have good performance with the exploration of regions where there is little information about the function’s response.

Source: A Bayesian Approach for the Robust Optimisation of Expensive-to-Evaluate Functions

Papers

Showing 211220 of 221 papers

TitleStatusHype
MEMe: An Accurate Maximum Entropy Method for Efficient Approximations in Large-Scale Machine Learning0
Meta-Learning surrogate models for sequential decision making0
Modelling the Effects of Hearing Loss on Neural Coding in the Auditory Midbrain with Variational Conditioning0
MONGOOSE: Path-wise Smooth Bayesian Optimisation via Meta-learning0
Mono-surrogate vs Multi-surrogate in Multi-objective Bayesian Optimisation0
Multi-fidelity Bayesian Optimisation of Syngas Fermentation Simulators0
Multi-fidelity Bayesian Optimisation with Continuous Approximations0
Multi-objective Bayesian optimisation with preferences over objectives0
Multi-view Bayesian optimisation in reduced dimension for engineering design0
On resampling vs. adjusting probabilistic graphical models in estimation of distribution algorithms0
Show:102550
← PrevPage 22 of 23Next →

No leaderboard results yet.