SOTAVerified

Fast and Accurate Neural CRF Constituency Parsing

2020-08-09IJCAI 2020Code Available1· sign in to hype

Yu Zhang, Houquan Zhou, Zhenghua Li

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Estimating probability distribution is one of the core issues in the NLP field. However, in both deep learning (DL) and pre-DL eras, unlike the vast applications of linear-chain CRF in sequence labeling tasks, very few works have applied tree-structure CRF to constituency parsing, mainly due to the complexity and inefficiency of the inside-outside algorithm. This work presents a fast and accurate neural CRF constituency parser. The key idea is to batchify the inside algorithm for loss computation by direct large tensor operations on GPU, and meanwhile avoid the outside algorithm for gradient computation via efficient back-propagation. We also propose a simple two-stage bracketing-then-labeling parsing approach to improve efficiency further. To improve the parsing performance, inspired by recent progress in dependency parsing, we introduce a new scoring architecture based on boundary representation and biaffine attention, and a beneficial dropout strategy. Experiments on PTB, CTB5.1, and CTB7 show that our two-stage CRF parser achieves new state-of-the-art performance on both settings of w/o and w/ BERT, and can parse over 1,000 sentences per second. We release our code at https://github.com/yzhangcs/crfpar.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CTB5CRF Parser + BERTF1 score92.27Unverified
CTB5CRF ParserF1 score89.8Unverified
CTB7CRF ParserF1 score88.6Unverified
CTB7CRF Parser + ElectraF1 score91.92Unverified
CTB7CRF Parser + BERTF1 score91.55Unverified
Penn TreebankCRF Parser + RoBERTaF1 score96.32Unverified
Penn TreebankCRF Parser + BERTF1 score95.69Unverified
Penn TreebankCRF ParserF1 score94.12Unverified

Reproductions