Useful nonrobust features are ubiquitous in biomedical images
Abstract
We study whether deep networks for medical imaging learn useful nonrobust features - predictive input patterns that are not human interpretable and highly susceptible to small adversarial perturbations - and how these features impact test performance. We show that models trained only on nonrobust features achieve well above chance accuracy across five MedMNIST classification tasks, confirming their predictive value in-distribution. Conversely, adversarially trained models that primarily rely on ...
Description / Details
We study whether deep networks for medical imaging learn useful nonrobust features - predictive input patterns that are not human interpretable and highly susceptible to small adversarial perturbations - and how these features impact test performance. We show that models trained only on nonrobust features achieve well above chance accuracy across five MedMNIST classification tasks, confirming their predictive value in-distribution. Conversely, adversarially trained models that primarily rely on robust features sacrifice in-distribution accuracy but yield markedly better performance under controlled distribution shifts (MedMNIST-C). Overall, nonrobust features boost standard accuracy yet degrade out-of-distribution performance, revealing a practical robustness-accuracy trade-off in medical imaging classification tasks that should be tailored to the requirements of the deployment setting.
Source: arXiv:2604.22579v1 - http://arxiv.org/abs/2604.22579v1 PDF: https://arxiv.org/pdf/2604.22579v1 Original Link: http://arxiv.org/abs/2604.22579v1
Please sign in to join the discussion.
No comments yet. Be the first to share your thoughts!
Apr 27, 2026
Biomedical Engineering
Engineering
0