SOTAVerified

Learning Activation Functions to Improve Deep Neural Networks

2014-12-21Code Available0· sign in to hype

Forest Agostinelli, Matthew Hoffman, Peter Sadowski, Pierre Baldi

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Artificial neural networks typically have a fixed, non-linear activation function at each neuron. We have designed a novel form of piecewise linear activation function that is learned independently for each neuron using gradient descent. With this adaptive activation function, we are able to improve upon deep neural network architectures composed of static rectified linear units, achieving state-of-the-art performance on CIFAR-10 (7.51%), CIFAR-100 (30.83%), and a benchmark from high-energy physics involving Higgs boson decay modes.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-10NiN+APLPercentage correct92.5Unverified
CIFAR-100NiN+APLPercentage correct69.2Unverified

Reproductions