SOTAVerified

Neural Interaction Transparency (NIT): Disentangling Learned Interactions for Improved Interpretability

2018-12-01NeurIPS 2018Unverified0· sign in to hype

Michael Tsang, Hanpeng Liu, Sanjay Purushotham, Pavankumar Murali, Yan Liu

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Neural networks are known to model statistical interactions, but they entangle the interactions at intermediate hidden layers for shared representation learning. We propose a framework, Neural Interaction Transparency (NIT), that disentangles the shared learning across different interactions to obtain their intrinsic lower-order and interpretable structure. This is done through a novel regularizer that directly penalizes interaction order. We show that disentangling interactions reduces a feedforward neural network to a generalized additive model with interactions, which can lead to transparent models that perform comparably to the state-of-the-art models. NIT is also flexible and efficient; it can learn generalized additive models with maximum K-order interactions by training only O(1) models.

Tasks

Reproductions