A Benchmark for Interpretability Methods in Deep Neural Networks
2018-06-28NeurIPS 2019Code Available0· sign in to hype
Sara Hooker, Dumitru Erhan, Pieter-Jan Kindermans, Been Kim
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/google-research/google-researchOfficialtf★ 37,519
- github.com/edpclau/ROAR-for-Scikitnone★ 0
- github.com/LLNL/fastcampytorch★ 0
Abstract
We propose an empirical measure of the approximate accuracy of feature importance estimates in deep neural networks. Our results across several large-scale image classification datasets show that many popular interpretability methods produce estimates of feature importance that are not better than a random designation of feature importance. Only certain ensemble based approaches---VarGrad and SmoothGrad-Squared---outperform such a random assignment of importance. The manner of ensembling remains critical, we show that some approaches do no better then the underlying method but carry a far higher computational burden.