SOTAVerified

CATT: Character-based Arabic Tashkeel Transformer

2024-07-03Code Available2· sign in to hype

Faris Alasmary, Orjuwan Zaafarani, Ahmad Ghannam

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Tashkeel, or Arabic Text Diacritization (ATD), greatly enhances the comprehension of Arabic text by removing ambiguity and minimizing the risk of misinterpretations caused by its absence. It plays a crucial role in improving Arabic text processing, particularly in applications such as text-to-speech and machine translation. This paper introduces a new approach to training ATD models. First, we finetuned two transformers, encoder-only and encoder-decoder, that were initialized from a pretrained character-based BERT. Then, we applied the Noisy-Student approach to boost the performance of the best model. We evaluated our models alongside 11 commercial and open-source models using two manually labeled benchmark datasets: WikiNews and our CATT dataset. Our findings show that our top model surpasses all evaluated models by relative Diacritic Error Rates (DERs) of 30.83\% and 35.21\% on WikiNews and CATT, respectively, achieving state-of-the-art in ATD. In addition, we show that our model outperforms GPT-4-turbo on CATT dataset by a relative DER of 9.36\%. We open-source our CATT models and benchmark dataset for the research communityhttps://github.com/abjadai/catt.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CATTCATT EDDER(%)8.62Unverified
CATTCATT EODER(%)8.76Unverified
CATTGPT-4DER(%)9.52Unverified
CATTCBHGDER(%)10.81Unverified
CATTCommand R+DER(%)13.17Unverified
CATTShakkalaDER(%)13.49Unverified
CATTSakhrDER(%)13.84Unverified
CATTAlkhalilDER(%)14.23Unverified
CATTMultilevel DiacritizerDER(%)16.48Unverified

Reproductions