SOTAVerified

Simplicial Attention Neural Networks

2022-03-14Code Available1· sign in to hype

L. Giusti, C. Battiloro, P. Di Lorenzo, S. Sardellitti, S. Barbarossa

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

The aim of this work is to introduce simplicial attention networks (SANs), i.e., novel neural architectures that operate on data defined on simplicial complexes leveraging masked self-attentional layers. Hinging on formal arguments from topological signal processing, we introduce a proper self-attention mechanism able to process data components at different layers (e.g., nodes, edges, triangles, and so on), while learning how to weight both upper and lower neighborhoods of the given topological domain in a totally task-oriented fashion. The proposed SANs generalize most of the current architectures available for processing data defined on simplicial complexes. The proposed approach compares favorably with other methods when applied to different (inductive and transductive) tasks such as trajectory prediction and missing data imputations in citation complexes.

Tasks

Reproductions