SOTAVerified

Pulling Out the Stops: Rethinking Stopword Removal for Topic Models

2017-04-01EACL 2017Unverified0· sign in to hype

Alex Schofield, ra, M{\aa}ns Magnusson, David Mimno

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

It is often assumed that topic models benefit from the use of a manually curated stopword list. Constructing this list is time-consuming and often subject to user judgments about what kinds of words are important to the model and the application. Although stopword removal clearly affects which word types appear as most probable terms in topics, we argue that this improvement is superficial, and that topic inference benefits little from the practice of removing stopwords beyond very frequent terms. Removing corpus-specific stopwords after model inference is more transparent and produces similar results to removing those words prior to inference.

Tasks

Reproductions