SOTAVerified

Bayesian Optimisation

Expensive black-box functions are a common problem in many disciplines, including tuning the parameters of machine learning algorithms, robotics, and other engineering design problems. Bayesian Optimisation is a principled and efficient technique for the global optimisation of these functions. The idea behind Bayesian Optimisation is to place a prior distribution over the target function and then update that prior with a set of “true” observations of the target function by expensively evaluating it in order to produce a posterior predictive distribution. The posterior then informs where to make the next observation of the target function through the use of an acquisition function, which balances the exploitation of regions known to have good performance with the exploration of regions where there is little information about the function’s response.

Source: A Bayesian Approach for the Robust Optimisation of Expensive-to-Evaluate Functions

Papers

Showing 141150 of 221 papers

TitleStatusHype
Uncertainty Aware System Identification with Universal Policies0
Uncovering Energy-Efficient Practices in Deep Learning Training: Preliminary Steps Towards Green AI0
Unsupervised machine learning for data-driven rock mass classification: addressing limitations in existing systems using drilling data0
What Makes an Effective Scalarising Function for Multi-Objective Bayesian Optimisation?0
Will More Expressive Graph Neural Networks do Better on Generative Tasks?0
Bayesian Search for Robust Optima0
Delayed Feedback in Kernel Bandits0
Detecting COVID-19 from Breathing and Coughing Sounds using Deep Neural Networks0
Developmental Bayesian Optimization of Black-Box with Visual Similarity-Based Transfer Learning0
Differential Evolution and Bayesian Optimisation for Hyper-Parameter Selection in Mixed-Signal Neuromorphic Circuits Applied to UAV Obstacle Avoidance0
Show:102550
← PrevPage 15 of 23Next →

No leaderboard results yet.