Probabilistic Rollouts for Learning Curve Extrapolation Across Hyperparameter Settings
2019-10-10Code Available0· sign in to hype
Matilde Gargiani, Aaron Klein, Stefan Falkner, Frank Hutter
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/gmatilde/vdrnnOfficialIn paperpytorch★ 0
Abstract
We propose probabilistic models that can extrapolate learning curves of iterative machine learning algorithms, such as stochastic gradient descent for training deep networks, based on training data with variable-length learning curves. We study instantiations of this framework based on random forests and Bayesian recurrent neural networks. Our experiments show that these models yield better predictions than state-of-the-art models from the hyperparameter optimization literature when extrapolating the performance of neural networks trained with different hyperparameter settings.