Universal Language Model Fine-tuning for Text Classification
Jeremy Howard, Sebastian Ruder
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/fastai/fastaiOfficialpytorch★ 27,931
- github.com/mrdbourke/tensorflow-deep-learningtf★ 5,869
- github.com/Deepayan137/Adapting-OCRpytorch★ 62
- github.com/tanmaylaud/Patient_Conversation_Classifier_FastAInone★ 1
- github.com/comicencyclo/TransferLearning_DiscriminativeFineTuningnone★ 1
- github.com/amagooda/SummaRuNNer_coattentionpytorch★ 1
- github.com/apmoore1/language-modelpytorch★ 1
- github.com/PrideLee/sentiment-analysispytorch★ 0
- github.com/castortroynz/desafio_atuacao19none★ 0
- github.com/Socialbird-AILab/BERT-Classification-Tutorialtf★ 0
Abstract
Inductive transfer learning has greatly impacted computer vision, but existing approaches in NLP still require task-specific modifications and training from scratch. We propose Universal Language Model Fine-tuning (ULMFiT), an effective transfer learning method that can be applied to any task in NLP, and introduce techniques that are key for fine-tuning a language model. Our method significantly outperforms the state-of-the-art on six text classification tasks, reducing the error by 18-24% on the majority of datasets. Furthermore, with only 100 labeled examples, it matches the performance of training from scratch on 100x more data. We open-source our pretrained models and code.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| IMDb | ULMFiT | Accuracy | 95.4 | — | Unverified |
| Yelp Binary classification | ULMFiT | Error | 2.16 | — | Unverified |
| Yelp Fine-grained classification | ULMFiT | Error | 29.98 | — | Unverified |