SOTAVerified

BenchIE: A Framework for Multi-Faceted Fact-Based Open Information Extraction Evaluation

2021-09-14ACL 2022Code Available1· sign in to hype

Kiril Gashteovski, Mingying Yu, Bhushan Kotnis, Carolin Lawrence, Mathias Niepert, Goran Glavaš

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Intrinsic evaluations of OIE systems are carried out either manually -- with human evaluators judging the correctness of extractions -- or automatically, on standardized benchmarks. The latter, while much more cost-effective, is less reliable, primarily because of the incompleteness of the existing OIE benchmarks: the ground truth extractions do not include all acceptable variants of the same fact, leading to unreliable assessment of the models' performance. Moreover, the existing OIE benchmarks are available for English only. In this work, we introduce BenchIE: a benchmark and evaluation framework for comprehensive evaluation of OIE systems for English, Chinese, and German. In contrast to existing OIE benchmarks, BenchIE is fact-based, i.e., it takes into account informational equivalence of extractions: our gold standard consists of fact synsets, clusters in which we exhaustively list all acceptable surface forms of the same fact. Moreover, having in mind common downstream applications for OIE, we make BenchIE multi-faceted; i.e., we create benchmark variants that focus on different facets of OIE evaluation, e.g., compactness or minimality of extractions. We benchmark several state-of-the-art OIE systems using BenchIE and demonstrate that these systems are significantly less effective than indicated by existing OIE benchmarks. We make BenchIE (data and evaluation code) publicly available on https://github.com/gkiril/benchie.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
BenchIEClausIEPrecision0.5Unverified
BenchIEMinIEPrecision0.43Unverified
BenchIEM2OIE (EN)Precision0.39Unverified
BenchIEROIE-TPrecision0.37Unverified
BenchIEOpenIE6Precision0.31Unverified
BenchIEM2OIE (ZH)Precision0.26Unverified
BenchIEROIE-NPrecision0.2Unverified
BenchIEStanford OIEPrecision0.11Unverified
BenchIEM2OIE (DE)Precision0.09Unverified
BenchIENaive OIEPrecision0.03Unverified

Reproductions