How Good Are My Predictions? Efficiently Approximating Precision-Recall Curves for Massive Datasets
Abstract
Large scale machine learning produces massive datasets whose items are often associated with a confidence level and can thus be ranked. However, computing the precision of these resources requires human annotation, which is often prohibitively expensive and is therefore skipped. We consider the problem of cost-effectively approximating precision-recall (PR) or ROC curves for such systems. Our novel approach, called PAULA, provides theoretically guaranteed lower and upper bounds on the underlying precision function while relying on only O(log N) annotations for a resource with N items. This contrasts favorably with Theta(sqrt(N \log N)) annotations needed by commonly used sampling based methods. Our key insight is to capitalize on a natural monotonicity property of the underlying confidence-based ranking. PAULA provides tight bounds for PR curves using, e.g., only 17K annotations for resources with 200K items and 48K annotations for resources with 2B items. We use PAULA to evaluate a subset of the much utilized PPDB paraphrase database and a recent Science knowledge base.
Cite
Text
Sabharwal and Sedghi. "How Good Are My Predictions? Efficiently Approximating Precision-Recall Curves for Massive Datasets." Conference on Uncertainty in Artificial Intelligence, 2017.Markdown
[Sabharwal and Sedghi. "How Good Are My Predictions? Efficiently Approximating Precision-Recall Curves for Massive Datasets." Conference on Uncertainty in Artificial Intelligence, 2017.](https://mlanthology.org/uai/2017/sabharwal2017uai-good/)BibTeX
@inproceedings{sabharwal2017uai-good,
title = {{How Good Are My Predictions? Efficiently Approximating Precision-Recall Curves for Massive Datasets}},
author = {Sabharwal, Ashish and Sedghi, Hanie},
booktitle = {Conference on Uncertainty in Artificial Intelligence},
year = {2017},
url = {https://mlanthology.org/uai/2017/sabharwal2017uai-good/}
}