A Mixture-of-Experts Model for Antonym-Synonym Discrimination
2021-08-01ACL 2021Code Available0· sign in to hype
Zhipeng Xie, Nan Zeng
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/zengnan1997/moe-asdOfficialIn paperpytorch★ 2
Abstract
Discrimination between antonyms and synonyms is an important and challenging NLP task. Antonyms and synonyms often share the same or similar contexts and thus are hard to make a distinction. This paper proposes two underlying hypotheses and employs the mixture-of-experts framework as a solution. It works on the basis of a divide-and-conquer strategy, where a number of localized experts focus on their own domains (or subspaces) to learn their specialties, and a gating mechanism determines the space partitioning and the expert mixture. Experimental results have shown that our method achieves the state-of-the-art performance on the task.