Unsupervised Hierarchical Story Infilling
Daphne Ippolito, David Grangier, Chris Callison-Burch, Douglas Eck
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Story infilling involves predicting words to go into a missing span from a story. This challenging task has the potential to transform interactive tools for creative writing. However, state-of-the-art conditional language models have trouble balancing fluency and coherence with novelty and diversity. We address this limitation with a hierarchical model which first selects a set of rare words and then generates text conditioned on that set. By relegating the high entropy task of picking rare words to a word-sampling model, the second-stage model conditioned on those words can achieve high fluency and coherence by searching for likely sentences, without sacrificing diversity.