SOTAVerified

Trainable Activations for Image Classification

2023-01-26Preprints 2023Code Available1· sign in to hype

Evgenii Pishchik

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Non-linear activation functions are one of the main parts of deep neural network architectures. The choice of the activation function can affect model speed, performance and convergence. Most popular activation functions don't have any trainable parameters and don't alter during the training. We propose different activation functions with and without trainable parameters. Said activation functions have a number of advantages and disadvantages. We'll be testing the performance of said activation functions and comparing the results with widely known activation function ReLU. We assume that the activation functions with trainable parameters can outperform functions without ones, because the trainable parameters allow the model to "select'' the type of each of the activation functions itself, however, this strongly depends on the architecture of the deep neural network and the activation function itself.

Tasks

Reproductions