SOTAVerified

Self-supervised Character-to-Character Distillation for Text Recognition

2022-11-01ICCV 2023Code Available1· sign in to hype

Tongkun Guan, Wei Shen, Xue Yang, Qi Feng, Zekun Jiang, Xiaokang Yang

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

When handling complicated text images (e.g., irregular structures, low resolution, heavy occlusion, and uneven illumination), existing supervised text recognition methods are data-hungry. Although these methods employ large-scale synthetic text images to reduce the dependence on annotated real images, the domain gap still limits the recognition performance. Therefore, exploring the robust text feature representations on unlabeled real images by self-supervised learning is a good solution. However, existing self-supervised text recognition methods conduct sequence-to-sequence representation learning by roughly splitting the visual features along the horizontal axis, which limits the flexibility of the augmentations, as large geometric-based augmentations may lead to sequence-to-sequence feature inconsistency. Motivated by this, we propose a novel self-supervised Character-to-Character Distillation method, CCD, which enables versatile augmentations to facilitate general text representation learning. Specifically, we delineate the character structures of unlabeled real images by designing a self-supervised character segmentation module. Following this, CCD easily enriches the diversity of local characters while keeping their pairwise alignment under flexible augmentations, using the transformation matrix between two augmented views from images. Experiments demonstrate that CCD achieves state-of-the-art results, with average performance gains of 1.38% in text recognition, 1.7% in text segmentation, 0.24 dB (PSNR) and 0.0321 (SSIM) in text super-resolution. Code is available at https://github.com/TongkunGuan/CCD.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CUTE80CCD-ViT-Small(ARD_2.8M)Accuracy98.3Unverified
CUTE80CCD-ViT-Tiny(ARD_2.8M)Accuracy95.8Unverified
CUTE80CCD-ViT-Base(ARD_2.8M)Accuracy98.3Unverified
HOSTCCD-ViT-Base1:1 Accuracy77.3Unverified
ICDAR2013CCD-ViT-Tiny(ARD_2.8M)Accuracy97.5Unverified
ICDAR2013CCD-ViT-Base(ARD_2.8M)Accuracy98.3Unverified
ICDAR2013CCD-ViT-Small(ARD_2.8M)Accuracy98.3Unverified
IIIT5kCCD-ViT-Tiny(ARD_2.8M)Accuracy97.1Unverified
IIIT5kCCD-ViT-Small(ARD_2.8M)Accuracy98Unverified
IIIT5kCCD-ViT-Base(ARD_2.8M)Accuracy98Unverified
SVTCCD-ViT-Base(ARD_2.8M)Accuracy97.8Unverified
SVTCCD-ViT-Tiny(ARD_2.8M)Accuracy96Unverified
SVTCCD-ViT-Small(ARD_2.8M)Accuracy96.4Unverified
SVTPCCD-ViT-BaseAccuracy96.1Unverified
SVTPCCD-ViT-SmallAccuracy92.7Unverified
SVTPCCD-ViT-TinyAccuracy91.6Unverified
WOSTCCD-ViT-Base1:1 Accuracy86Unverified

Reproductions