Class Prior Estimation with Biased Positives and Unlabeled Examples
release_gu2dji4nprgtbpsmp5lljbntai
by
Shantanu Jain, Justin Delano, Himanshu Sharma, Predrag Radivojac
2020 Volume 34, Issue 04, p4255-4263
Abstract
Positive-unlabeled learning is often studied under the assumption that the labeled positive sample is drawn randomly from the true distribution of positives. In many application domains, however, certain regions in the support of the positive class-conditional distribution are over-represented while others are under-represented in the positive sample. Although this introduces problems in all aspects of positive-unlabeled learning, we begin to address this challenge by focusing on the estimation of class priors, quantities central to the estimation of posterior probabilities and the recovery of true classification performance. We start by making a set of assumptions to model the sampling bias. We then extend the identifiability theory of class priors from the unbiased to the biased setting. Finally, we derive an algorithm for estimating the class priors that relies on clustering to decompose the original problem into subproblems of unbiased positive-unlabeled learning. Our empirical investigation suggests feasibility of the correction strategy and overall good performance.
In application/xml+jats
format
Archived Files and Locations
application/pdf 592.6 kB
file_se6ls6ezyvfkpbqvkft6oqvxmq
|
aaai.org (web) web.archive.org (webarchive) |
article-journal
Stage
published
Date 2020-04-03
access all versions, variants, and formats of this works (eg, pre-prints)
Crossref Metadata (via API)
Worldcat
SHERPA/RoMEO (journal policies)
wikidata.org
CORE.ac.uk
Semantic Scholar
Google Scholar