SOTAVerified

TAMER: Tree-Aware Transformer for Handwritten Mathematical Expression Recognition

2024-08-16Code Available1· sign in to hype

Jianhua Zhu, Wenqi Zhao, Yu Li, Xingjian Hu, Liangcai Gao

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Handwritten Mathematical Expression Recognition (HMER) has extensive applications in automated grading and office automation. However, existing sequence-based decoding methods, which directly predict sequences, struggle to understand and model the inherent tree structure of and often fail to ensure syntactic correctness in the decoded results. To address these challenges, we propose a novel model named TAMER (Tree-Aware Transformer) for handwritten mathematical expression recognition. TAMER introduces an innovative Tree-aware Module while maintaining the flexibility and efficient training of Transformer. TAMER combines the advantages of both sequence decoding and tree decoding models by jointly optimizing sequence prediction and tree structure prediction tasks, which enhances the model's understanding and generalization of complex mathematical expression structures. During inference, TAMER employs a Tree Structure Prediction Scoring Mechanism to improve the structural validity of the generated sequences. Experimental results on CROHME datasets demonstrate that TAMER outperforms traditional sequence decoding and tree decoding models, especially in handling complex mathematical structures, achieving state-of-the-art (SOTA) performance.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CROHME 2014TAMERExpRate61.23Unverified
CROHME 2016TAMERExpRate60.26Unverified
CROHME 2019TAMERExpRate61.97Unverified
HME100KTAMERExpRate68.52Unverified

Reproductions