Publication

High Probability Risk Control Under Covariate Shift

Duarte C. Almeida, João Bravo, Jacopo Bono, Pedro Bizarro, Mário A. T. Figueiredo

Published at COPA 2025

AI Research

Abstract

Distribution-free uncertainty quantification is an emerging field, which encompasses risk control techniques in finite sample settings with minimal distributional assumptions, making it suitable for high-stakes applications. In particular, high-probability risk control methods, namely the learn then test (LTT) framework, use a calibration set to control multiple risks with high confidence. However, these methods rely on the assumption that the calibration and target distributions are identical, which can pose challenges, for example, when controlling label-dependent risks under the absence of labeled target data. In this work, we propose a novel extension of LTT that handles covariate shifts by directly weighting calibration losses with importance weights. We validate our method on a synthetic fraud detection task, aiming to control the false positive rate while minimizing false negatives, and on an image classification task, to control the miscoverage of a set predictor while minimizing the average set size. The results show that our approach consistently yields less conservative risk control than existing baselines based on rejection sampling, which results in overall lower false negative rates and smaller prediction sets.

Materials
PDF

Page printed in 4 Oct 2025. Plase see https://research.feedzai.com/publication/high-probability-risk-control-under-covariate-shift for the latest version.