SOTAVerified

Randomized Ablation Feature Importance

2019-10-01Unverified0· sign in to hype

Luke Merrick

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Given a model f that predicts a target y from a vector of input features x = x_1, x_2, , x_M, we seek to measure the importance of each feature with respect to the model's ability to make a good prediction. To this end, we consider how (on average) some measure of goodness or badness of prediction (which we term "loss" ), changes when we hide or ablate each feature from the model. To ablate a feature, we replace its value with another possible value randomly. By averaging over many points and many possible replacements, we measure the importance of a feature on the model's ability to make good predictions. Furthermore, we present statistical measures of uncertainty that quantify how confident we are that the feature importance we measure from our finite dataset and finite number of ablations is close to the theoretical true importance value.

Tasks

Reproductions