SOTAVerified

Understanding Isomorphism Bias in Graph Data Sets

2019-10-26Code Available0· sign in to hype

Sergei Ivanov, Sergei Sviridov, Evgeny Burnaev

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

In recent years there has been a rapid increase in classification methods on graph structured data. Both in graph kernels and graph neural networks, one of the implicit assumptions of successful state-of-the-art models was that incorporating graph isomorphism features into the architecture leads to better empirical performance. However, as we discover in this work, commonly used data sets for graph classification have repeating instances which cause the problem of isomorphism bias, i.e. artificially increasing the accuracy of the models by memorizing target information from the training set. This prevents fair competition of the algorithms and raises a question of the validity of the obtained results. We analyze 54 data sets, previously extensively used for graph-related tasks, on the existence of isomorphism bias, give a set of recommendations to machine learning practitioners to properly set up their models, and open source new data sets for the future experiments.

Tasks

Reproductions