Learning Deep Features for Discriminative Localization
Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, Antonio Torralba
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/zhoubolei/CAMOfficialtf★ 0
- github.com/frgfm/torch-campytorch★ 2,291
- github.com/joaopauloschuler/k-neural-apitf★ 135
- github.com/Azure/AzureChestXRaypytorch★ 88
- github.com/windstormer/Cfd-CAMpytorch★ 3
- github.com/jsr66/Machine-Learning-Phases-of-Matter-with-Discriminative-Localization-none★ 0
- github.com/zdcuob/Fully-Convlutional-Neural-Networks-for-state-of-the-art-time-series-classification-tf★ 0
- github.com/jsr66/Machine-Learning-Phases-of-Matter-with-Discriminative-Localizationnone★ 0
- github.com/Seb-Good/deepecgtf★ 0
- github.com/HRanWang/Spatial-Re-Scalingpytorch★ 0
Abstract
In this work, we revisit the global average pooling layer proposed in [13], and shed light on how it explicitly enables the convolutional neural network to have remarkable localization ability despite being trained on image-level labels. While this technique was previously proposed as a means for regularizing training, we find that it actually builds a generic localizable deep representation that can be applied to a variety of tasks. Despite the apparent simplicity of global average pooling, we are able to achieve 37.1% top-5 error for object localization on ILSVRC 2014, which is remarkably close to the 34.2% top-5 error achieved by a fully supervised CNN approach. We demonstrate that our network is able to localize the discriminative image regions on a variety of tasks despite not being trained for them