SOTAVerified

ADADELTA: An Adaptive Learning Rate Method

2012-12-22Code Available1· sign in to hype

Matthew D. Zeiler

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We present a novel per-dimension learning rate method for gradient descent called ADADELTA. The method dynamically adapts over time using only first order information and has minimal computational overhead beyond vanilla stochastic gradient descent. The method requires no manual tuning of a learning rate and appears robust to noisy gradient information, different model architecture choices, various data modalities and selection of hyperparameters. We show promising results compared to other methods on the MNIST digit classification task using a single machine and on a large scale voice dataset in a distributed cluster environment.

Tasks

Reproductions