SOTAVerified

Tree-Structured Topic Modeling with Nonparametric Neural Variational Inference

2021-08-01ACL 2021Code Available1· sign in to hype

Ziye Chen, Cheng Ding, Zusheng Zhang, Yanghui Rao, Haoran Xie

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Topic modeling has been widely used for discovering the latent semantic structure of documents, but most existing methods learn topics with a flat structure. Although probabilistic models can generate topic hierarchies by introducing nonparametric priors like Chinese restaurant process, such methods have data scalability issues. In this study, we develop a tree-structured topic model by leveraging nonparametric neural variational inference. Particularly, the latent components of the stick-breaking process are first learned for each document, then the affiliations of latent components are modeled by the dependency matrices between network layers. Utilizing this network structure, we can efficiently extract a tree-structured topic hierarchy with reasonable structure, low redundancy, and adaptable widths. Experiments on real-world datasets validate the effectiveness of our method.

Tasks

Reproductions