Improving Korean NLP Tasks with Linguistically Informed Subword Tokenization and Sub-character Decomposition
Taehee Jeon, BongSeok Yang, ChangHwan Kim, Yoonseob Lim
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/taeheejeon22/morphsubdecomp-koreanOfficialIn paperpytorch★ 4
Abstract
We introduce a morpheme-aware subword tokenization method that utilizes sub-character decomposition to address the challenges of applying Byte Pair Encoding (BPE) to Korean, a language characterized by its rich morphology and unique writing system. Our approach balances linguistic accuracy with computational efficiency in Pre-trained Language Models (PLMs). Our evaluations show that this technique achieves good performances overall, notably improving results in the syntactic task of NIKL-CoLA. This suggests that integrating morpheme type information can enhance language models' syntactic and semantic capabilities, indicating that adopting more linguistic insights can further improve performance beyond standard morphological analysis.