SOTAVerified

A Protocol for Continual Explanation of SHAP

2023-06-12Code Available0· sign in to hype

Andrea Cossu, Francesco Spinnato, Riccardo Guidotti, Davide Bacciu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Continual Learning trains models on a stream of data, with the aim of learning new information without forgetting previous knowledge. Given the dynamic nature of such environments, explaining the predictions of these models can be challenging. We study the behavior of SHAP values explanations in Continual Learning and propose an evaluation protocol to robustly assess the change of explanations in Class-Incremental scenarios. We observed that, while Replay strategies enforce the stability of SHAP values in feedforward/convolutional models, they are not able to do the same with fully-trained recurrent models. We show that alternative recurrent approaches, like randomized recurrent models, are more effective in keeping the explanations stable over time.

Tasks

Reproductions