On Continuous Monitoring of Risk Violations under Unknown Shift

Alexander Timans, Rajeev Verma, Eric Nalisnick, Christian A. Naesseth
Proceedings of the Forty-first Conference on Uncertainty in Artificial Intelligence, PMLR 286:4204-4226, 2025.

Abstract

Machine learning systems deployed in the real world must operate under dynamic and often unpredictable distribution shifts. This challenges the validity of statistical safety assurances on the system’s risk established beforehand. Common risk control frameworks rely on fixed assumptions and lack mechanisms to continuously monitor deployment reliability. In this work, we propose a general framework for the real-time monitoring of risk violations in evolving data streams. Leveraging the ‘testing by betting’ paradigm, we propose a sequential hypothesis testing procedure to detect violations of bounded risks associated with the model’s decision-making mechanism, while ensuring control on the false alarm rate. Our method operates under minimal assumptions on the nature of encountered shifts, rendering it broadly applicable. We illustrate the effectiveness of our approach by monitoring risks in outlier detection and set prediction under a variety of shifts.

Cite this Paper


BibTeX
@InProceedings{pmlr-v286-timans25a, title = {On Continuous Monitoring of Risk Violations under Unknown Shift}, author = {Timans, Alexander and Verma, Rajeev and Nalisnick, Eric and Naesseth, Christian A.}, booktitle = {Proceedings of the Forty-first Conference on Uncertainty in Artificial Intelligence}, pages = {4204--4226}, year = {2025}, editor = {Chiappa, Silvia and Magliacane, Sara}, volume = {286}, series = {Proceedings of Machine Learning Research}, month = {21--25 Jul}, publisher = {PMLR}, pdf = {https://raw.githubusercontent.com/mlresearch/v286/main/assets/timans25a/timans25a.pdf}, url = {https://proceedings.mlr.press/v286/timans25a.html}, abstract = {Machine learning systems deployed in the real world must operate under dynamic and often unpredictable distribution shifts. This challenges the validity of statistical safety assurances on the system’s risk established beforehand. Common risk control frameworks rely on fixed assumptions and lack mechanisms to continuously monitor deployment reliability. In this work, we propose a general framework for the real-time monitoring of risk violations in evolving data streams. Leveraging the ‘testing by betting’ paradigm, we propose a sequential hypothesis testing procedure to detect violations of bounded risks associated with the model’s decision-making mechanism, while ensuring control on the false alarm rate. Our method operates under minimal assumptions on the nature of encountered shifts, rendering it broadly applicable. We illustrate the effectiveness of our approach by monitoring risks in outlier detection and set prediction under a variety of shifts.} }
Endnote
%0 Conference Paper %T On Continuous Monitoring of Risk Violations under Unknown Shift %A Alexander Timans %A Rajeev Verma %A Eric Nalisnick %A Christian A. Naesseth %B Proceedings of the Forty-first Conference on Uncertainty in Artificial Intelligence %C Proceedings of Machine Learning Research %D 2025 %E Silvia Chiappa %E Sara Magliacane %F pmlr-v286-timans25a %I PMLR %P 4204--4226 %U https://proceedings.mlr.press/v286/timans25a.html %V 286 %X Machine learning systems deployed in the real world must operate under dynamic and often unpredictable distribution shifts. This challenges the validity of statistical safety assurances on the system’s risk established beforehand. Common risk control frameworks rely on fixed assumptions and lack mechanisms to continuously monitor deployment reliability. In this work, we propose a general framework for the real-time monitoring of risk violations in evolving data streams. Leveraging the ‘testing by betting’ paradigm, we propose a sequential hypothesis testing procedure to detect violations of bounded risks associated with the model’s decision-making mechanism, while ensuring control on the false alarm rate. Our method operates under minimal assumptions on the nature of encountered shifts, rendering it broadly applicable. We illustrate the effectiveness of our approach by monitoring risks in outlier detection and set prediction under a variety of shifts.
APA
Timans, A., Verma, R., Nalisnick, E. & Naesseth, C.A.. (2025). On Continuous Monitoring of Risk Violations under Unknown Shift. Proceedings of the Forty-first Conference on Uncertainty in Artificial Intelligence, in Proceedings of Machine Learning Research 286:4204-4226 Available from https://proceedings.mlr.press/v286/timans25a.html.

Related Material