Bias Reduction through Conditional Conformal Prediction
2015 (English)In: Intelligent Data Analysis, ISSN 1088-467X, E-ISSN 1571-4128, Vol. 9, no 6, 1355-1375 p.Article in journal (Refereed) Published
Conformal prediction (CP) is a relatively new framework in which predictive models output sets of predictions with a bound on the error rate, i.e., the probability of making an erroneous prediction is guaranteed to be equal to or less than a predefined significance level. Label-conditional conformal prediction (LCCP) is a specialization of the framework which gives a bound on the error rate for each individual class. For datasets with class imbalance, many learning algorithms have a tendency to predict the majority class more often than the expected relative frequency, i.e., they are biased in favor of the majority class. In this study, the class bias of standard and label-conditional conformal predictors is investigated. An empirical investigation on 32 publicly available datasets with varying degrees of class imbalance is presented. The experimental results show that CP is highly biased towards the majority class on imbalanced datasets, i.e., it can be expected to make a majority of its errors on the minority class. LCCP, on the other hand, is not biased towards the majority class. Instead, the errors are distributed between the classes almost in accordance with the prior class distribution.
Place, publisher, year, edition, pages
2015. Vol. 9, no 6, 1355-1375 p.
Conformal prediction, imbalanced learning, class bias
Research subject Computer and Systems Sciences
IdentifiersURN: urn:nbn:se:su:diva-116682DOI: 10.3233/IDA-150786ISI: 000366058000010OAI: oai:DiVA.org:su-116682DiVA: diva2:807229