SOTAVerified

SoftAdam: Unifying SGD and Adam for better stochastic gradient descent

2019-09-25Unverified0· sign in to hype

Abraham J. Fetterman, Christina H. Kim, Joshua Albrecht

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Abstract Stochastic gradient descent (SGD) and Adam are commonly used to optimize deep neural networks, but choosing one usually means making tradeoffs between speed, accuracy and stability. Here we present an intuition for why the tradeoffs exist as well as a method for unifying the two in a continuous way. This makes it possible to control the way models are trained in much greater detail. We show that for default parameters, the new algorithm equals or outperforms SGD and Adam across a range of models for image classification tasks and outperforms SGD for language modeling tasks.

Tasks

Reproductions