SOTAVerified

Bayesian Optimisation

Expensive black-box functions are a common problem in many disciplines, including tuning the parameters of machine learning algorithms, robotics, and other engineering design problems. Bayesian Optimisation is a principled and efficient technique for the global optimisation of these functions. The idea behind Bayesian Optimisation is to place a prior distribution over the target function and then update that prior with a set of “true” observations of the target function by expensively evaluating it in order to produce a posterior predictive distribution. The posterior then informs where to make the next observation of the target function through the use of an acquisition function, which balances the exploitation of regions known to have good performance with the exploration of regions where there is little information about the function’s response.

Source: A Bayesian Approach for the Robust Optimisation of Expensive-to-Evaluate Functions

Papers

Showing 2130 of 221 papers

TitleStatusHype
Developing Optimal Causal Cyber-Defence Agents via Cyber Security SimulationCode1
Learning to Do or Learning While Doing: Reinforcement Learning and Bayesian Optimisation for Online Continuous TuningCode1
AutoLRS: Automatic Learning-Rate Schedule by Bayesian Optimization on the FlyCode1
Applications of Gaussian Processes at Extreme Lengthscales: From Molecules to Black HolesCode1
Adversarial Attacks on Graph Classifiers via Bayesian OptimisationCode1
Neural Diffusion ProcessesCode1
Neural Architecture Generator OptimizationCode1
SOBER: Highly Parallel Bayesian Optimization and Bayesian Quadrature over Discrete and Mixed SpacesCode1
Approximate Neural Architecture Search via Operation Distribution Learning0
Automatic Tuning of Stochastic Gradient Descent with Bayesian Optimisation0
Show:102550
← PrevPage 3 of 23Next →

No leaderboard results yet.