SOTAVerified

Precision-Recall-Gain Curves: PR Analysis Done Right

2015-12-01NeurIPS 2015Code Available0· sign in to hype

Peter Flach, Meelis Kull

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Precision-Recall analysis abounds in applications of binary classification where true negatives do not add value and hence should not affect assessment of the classifier's performance. Perhaps inspired by the many advantages of receiver operating characteristic (ROC) curves and the area under such curves for accuracy-based performance assessment, many researchers have taken to report Precision-Recall (PR) curves and associated areas as performance metric. We demonstrate in this paper that this practice is fraught with difficulties, mainly because of incoherent scale assumptions -- e.g., the area under a PR curve takes the arithmetic mean of precision values whereas the F_ score applies the harmonic mean. We show how to fix this by plotting PR curves in a different coordinate system, and demonstrate that the new Precision-Recall-Gain curves inherit all key advantages of ROC curves. In particular, the area under Precision-Recall-Gain curves conveys an expected F_1 score on a harmonic scale, and the convex hull of a Precision-Recall-Gain curve allows us to calibrate the classifier's scores so as to determine, for each operating point on the convex hull, the interval of values for which the point optimises F_. We demonstrate experimentally that the area under traditional PR curves can easily favour models with lower expected F_1 score than others, and so the use of Precision-Recall-Gain curves will result in better model selection.

Tasks

Reproductions