SOTAVerified

HUBERT Untangles BERT to Improve Transfer across NLP Tasks

2019-10-25Code Available0· sign in to hype

Mehrad Moradshahi, Hamid Palangi, Monica S. Lam, Paul Smolensky, Jianfeng Gao

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We introduce HUBERT which combines the structured-representational power of Tensor-Product Representations (TPRs) and BERT, a pre-trained bidirectional Transformer language model. We show that there is shared structure between different NLP datasets that HUBERT, but not BERT, is able to learn and leverage. We validate the effectiveness of our model on the GLUE benchmark and HANS dataset. Our experiment results show that untangling data-specific semantics from general language structure is key for better transfer among NLP tasks.

Tasks

Reproductions