Back to Explorer
Research PaperResearchia:202603.25056[Data Science > Machine Learning]

Robustness Quantification for Discriminative Models: a New Robustness Metric and its Application to Dynamic Classifier Selection

Rodrigo F. L. Lassance

Abstract

Among the different possible strategies for evaluating the reliability of individual predictions of classifiers, robustness quantification stands out as a method that evaluates how much uncertainty a classifier could cope with before changing its prediction. However, its applicability is more limited than some of its alternatives, since it requires the use of generative models and restricts the analyses either to specific model architectures or discrete features. In this work, we propose a new robustness metric applicable to any probabilistic discriminative classifier and any type of features. We demonstrate that this new metric is capable of distinguishing between reliable and unreliable predictions, and use this observation to develop new strategies for dynamic classifier selection.


Source: arXiv:2603.23318v1 - http://arxiv.org/abs/2603.23318v1 PDF: https://arxiv.org/pdf/2603.23318v1 Original Link: http://arxiv.org/abs/2603.23318v1

Submission:3/25/2026
Comments:0 comments
Subjects:Machine Learning; Data Science
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Robustness Quantification for Discriminative Models: a New Robustness Metric and its Application to Dynamic Classifier Selection | Researchia