SOTAVerified

Empirical Evaluation of Rectified Activations in Convolutional Network

2015-05-05Code Available0· sign in to hype

Bing Xu, Naiyan Wang, Tianqi Chen, Mu Li

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In this paper we investigate the performance of different types of rectified activation functions in convolutional neural network: standard rectified linear unit (ReLU), leaky rectified linear unit (Leaky ReLU), parametric rectified linear unit (PReLU) and a new randomized leaky rectified linear units (RReLU). We evaluate these activation function on standard image classification task. Our experiments suggest that incorporating a non-zero slope for negative part in rectified activation units could consistently improve the results. Thus our findings are negative on the common belief that sparsity is the key of good performance in ReLU. Moreover, on small scale dataset, using deterministic negative slope or learning it are both prone to overfitting. They are not as effective as using their randomized counterpart. By using RReLU, we achieved 75.68\% accuracy on CIFAR-100 test set without multiple test or ensemble.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CIFAR-10RReLUPercentage correct88.8Unverified
CIFAR-100RReLUPercentage correct59.8Unverified

Reproductions