SOTAVerified

Recovering Barabási-Albert Parameters of Graphs through Disentanglement

2021-05-03ICLR Workshop GTRL 2021Code Available0· sign in to hype

Cristina Guzman, Daphna Keidar, Tristan Meynier, Andreas Opedal, Niklas Stoehr

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Classical graph modeling approaches such as Erdos R\'enyi (ER) random graphs or Barab\'asi-Albert (BA) graphs, here referred to as stylized models, aim to reproduce properties of real-world graphs in an interpretable way. While useful, graph generation with stylized models requires domain knowledge and iterative trial and error simulation. Previous work by Stoehr et al. (2019) addresses these issues by learning the generation process from graph data, using a disentanglement-focused deep autoencoding framework, more specifically, a -Variational Autoencoder (-VAE). While they successfully recover the generative parameters of ER graphs through the model's latent variables, their model performs badly on sequentially generated graphs such as BA graphs, due to their oversimplified decoder. We focus on recovering the generative parameters of BA graphs by replacing their -VAE decoder with a sequential one. We first learn the generative BA parameters in a supervised fashion using a Graph Neural Network (GNN) and a Random Forest Regressor, by minimizing the squared loss between the true generative parameters and the latent variables. Next, we train a -VAE model, combining the GNN encoder from the first stage with an LSTM-based decoder with a customized loss.

Tasks

Reproductions