SOTAVerified

Persformer: A Transformer Architecture for Topological Machine Learning

2021-12-30Code Available1· sign in to hype

Raphael Reinauer, Matteo Caorsi, Nicolas Berkouk

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

One of the main challenges of Topological Data Analysis (TDA) is to extract features from persistent diagrams directly usable by machine learning algorithms. Indeed, persistence diagrams are intrinsically (multi-)sets of points in R^2 and cannot be seen in a straightforward manner as vectors. In this article, we introduce Persformer, the first Transformer neural network architecture that accepts persistence diagrams as input. The Persformer architecture significantly outperforms previous topological neural network architectures on classical synthetic and graph benchmark datasets. Moreover, it satisfies a universal approximation theorem. This allows us to introduce the first interpretability method for topological machine learning, which we explore in two examples.

Tasks

Reproductions