Skip-Thought Vectors
Ryan Kiros, Yukun Zhu, Ruslan Salakhutdinov, Richard S. Zemel, Antonio Torralba, Raquel Urtasun, Sanja Fidler
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/facebookresearch/InferSentpytorch★ 2,280
- github.com/soskek/bookcorpusnone★ 852
- github.com/dashayushman/TAC-GANtf★ 105
- github.com/kushalpatil1997/text_to_image_synthesistf★ 2
- github.com/dwright37/phylogenetic-autoencodertf★ 0
- github.com/thomasyue/tf2-skip-thoughtstf★ 0
- github.com/whitneysattler/Skip-Thoughtsnone★ 0
- github.com/luweizhang/joint_embeddingspytorch★ 0
- github.com/arukavina/baking-lyricstf★ 0
- github.com/ryankiros/skip-thoughtsnone★ 0
Abstract
We describe an approach for unsupervised learning of a generic, distributed sentence encoder. Using the continuity of text from books, we train an encoder-decoder model that tries to reconstruct the surrounding sentences of an encoded passage. Sentences that share semantic and syntactic properties are thus mapped to similar vector representations. We next introduce a simple vocabulary expansion method to encode words that were not seen as part of training, allowing us to expand our vocabulary to a million words. After training our model, we extract and evaluate our vectors with linear models on 8 tasks: semantic relatedness, paraphrase detection, image-sentence ranking, question-type classification and 4 benchmark sentiment and subjectivity datasets. The end result is an off-the-shelf encoder that can produce highly generic sentence representations that are robust and perform well in practice. We will make our encoder publicly available.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| SICK | combine-skip (Kiros et al., 2015) | MSE | 0.27 | — | Unverified |