Striving for Simplicity: The All Convolutional Net
Jost Tobias Springenberg, Alexey Dosovitskiy, Thomas Brox, Martin Riedmiller
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/billyshin/CNN-Object-Recognitiontf★ 1
- github.com/AlexSzen/wwf_shark_idpytorch★ 0
- github.com/imavijit/detect_itnone★ 0
- github.com/BasileBron/Artificial-Intelligence-Lexicalnone★ 0
- github.com/hs2k/pytorch-smoothgradpytorch★ 0
- github.com/axit54/Object-Recognition-using-the-All-CNN-networktf★ 0
- github.com/eclique/keras-gradcamtf★ 0
- github.com/balshersingh10/Object-Recognitionnone★ 0
- github.com/vpaliwal1/Image-Classification-using-CNNtf★ 0
- github.com/Paperspace/hyperopt-keras-sampletf★ 0
Abstract
Most modern convolutional neural networks (CNNs) used for object recognition are built using the same principles: Alternating convolution and max-pooling layers followed by a small number of fully connected layers. We re-evaluate the state of the art for object recognition from small images with convolutional networks, questioning the necessity of different components in the pipeline. We find that max-pooling can simply be replaced by a convolutional layer with increased stride without loss in accuracy on several image recognition benchmarks. Following this finding -- and building on other recent work for finding simple network structures -- we propose a new architecture that consists solely of convolutional layers and yields competitive or state of the art performance on several object recognition datasets (CIFAR-10, CIFAR-100, ImageNet). To analyze the network we introduce a new variant of the "deconvolution approach" for visualizing features learned by CNNs, which can be applied to a broader range of network structures than existing approaches.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| CIFAR-10 | ACN | Percentage correct | 95.6 | — | Unverified |
| CIFAR-100 | ACN | Percentage correct | 66.3 | — | Unverified |