Useful nonrobust features are ubiquitous in biomedical images

TL;DR

The study shows that deep networks learn useful nonrobust features in medical images, achieving high accuracy on five MedMNIST tasks.

eess.IV 🔴 Advanced 2026-04-24 16 views
Coenraad Mouton Randle Rabe Niklas C. Koser Nicolai Krekiehn Christopher Hansen Jan-Bernd Hövener Claus-C. Glüer
medical imaging deep learning nonrobust features adversarial training out-of-distribution performance

Key Findings

Methodology

The study employs deep neural networks (DNNs) to classify medical images, investigating whether they learn useful nonrobust features. Experiments on five MedMNIST datasets confirm the predictive value of nonrobust features. Using adversarial training, models relying primarily on robust features perform better in out-of-distribution tests, while nonrobust features enhance standard accuracy.

Key Results

  • On five MedMNIST datasets, models relying solely on nonrobust features perform better than random guessing in in-distribution tests, demonstrating their predictive value. For instance, the nonrobust model achieves a balanced accuracy of 62% on PathMNIST.
  • Adversarially trained models perform excellently under controlled distribution shifts (MedMNIST-C), with balanced accuracy increasing from 78% to 89% on PathMNIST.
  • Nonrobust feature models perform poorly under adversarial attacks, with accuracy dropping to 4%, whereas adversarially trained models perform significantly better, with accuracy ranging from 57% to 73%.

Significance

The study reveals a robustness-accuracy trade-off in medical imaging classification tasks. While nonrobust features improve standard accuracy, they perform poorly in out-of-distribution tests. This finding is significant for model deployment in medical imaging analysis, suggesting the need to balance robustness and accuracy in different application scenarios.

Technical Contribution

This study provides the first systematic investigation of the generalization properties of robust and nonrobust features in medical image classification tasks. By using adversarial training, it demonstrates the advantage of robust features in out-of-distribution tests and highlights the potential risks of applying nonrobust features in medical imaging.

Novelty

This is the first systematic study of nonrobust features in the medical imaging domain. Compared to previous studies on natural images, this paper focuses on medical images, revealing the widespread presence and impact of nonrobust features in this field.

Limitations

  • The study primarily focuses on two-dimensional low-resolution images, not involving three-dimensional or high-resolution imaging, which may limit the generalizability of the conclusions.
  • The evaluation does not cover natural distribution shifts; future studies should consider independently sourced datasets.
  • The experiments are limited to CNNs, and the performance of modern architectures like vision transformers has not been verified.

Future Work

Future research could extend to higher-resolution or three-dimensional imaging, explore the performance of modern architectures like vision transformers, and focus on the impact of natural distribution shifts. Additionally, studies should verify these findings in segmentation or detection tasks.

AI Executive Summary

In medical imaging analysis, deep neural networks (DNNs) are increasingly applied across domains such as radiology, digital pathology, and ophthalmology. However, whether these models learn useful features, particularly nonrobust features, remains an open question. Nonrobust features are input patterns highly susceptible to small adversarial perturbations and difficult to interpret. While extensively studied in natural images, their role in medical imaging is not well understood.

This study demonstrates that deep networks learn useful nonrobust features in medical images, achieving high accuracy on five MedMNIST classification tasks. Experiments reveal that models relying solely on nonrobust features perform better than random guessing in in-distribution tests, showcasing their predictive value. However, these features perform poorly in out-of-distribution tests, especially under adversarial attacks, where accuracy significantly drops.

To investigate the role of robust features, the authors employ adversarial training. Models relying primarily on robust features perform better under controlled distribution shifts (MedMNIST-C), indicating that robust features offer better stability and generalization when facing distribution changes.

The findings reveal a robustness-accuracy trade-off in medical imaging classification tasks. While nonrobust features improve standard accuracy, they perform poorly in out-of-distribution tests. This discovery is significant for model deployment in medical imaging analysis, suggesting the need to balance robustness and accuracy in different application scenarios.

However, the study has some limitations. It primarily focuses on two-dimensional low-resolution images, not involving three-dimensional or high-resolution imaging, which may limit the generalizability of the conclusions. Additionally, the evaluation does not cover natural distribution shifts, and future studies should consider independently sourced datasets.

In conclusion, this paper offers a new perspective on feature learning in medical imaging analysis, emphasizing the importance of considering robustness in model development and deployment. Future research could extend to higher-resolution or three-dimensional imaging, explore the performance of modern architectures like vision transformers, and verify these findings in segmentation or detection tasks.

Deep Analysis

Background

Medical imaging analysis is a crucial application area for artificial intelligence, involving fields such as radiology, digital pathology, and ophthalmology. With the advancement of deep learning technology, deep neural networks (DNNs) are increasingly applied in these fields. However, the reliability, interpretability, and robustness of these models remain critical issues. Previous studies on natural images have shown that neural networks tend to learn nonrobust features, which are highly sensitive to small perturbations and difficult to interpret. The presence and impact of these features in medical imaging have not been fully explored. This paper aims to systematically investigate the role of robust and nonrobust features in medical imaging, particularly their performance in in-distribution and out-of-distribution tests.

Core Problem

The core problem in applying deep neural networks to medical imaging analysis is whether the models learn useful nonrobust features. While these features perform well in in-distribution tests, they may lead to performance degradation in out-of-distribution tests. Nonrobust features are highly sensitive to small adversarial perturbations and difficult to interpret, potentially increasing the model's susceptibility to adversarial examples. Investigating the role of these features in medical imaging is crucial for improving model robustness and interpretability.

Innovation

The core innovations of this paper include the first systematic investigation of the generalization properties of robust and nonrobust features in medical image classification tasks. • Experiments on five MedMNIST datasets confirm the predictive value of nonrobust features. • Using adversarial training, models relying primarily on robust features perform better in out-of-distribution tests. • The study highlights the potential risks of applying nonrobust features in medical imaging, emphasizing the importance of considering robustness in model development and deployment.

Methodology

The study employs the following methods:


  • �� Dataset Selection: Five MedMNIST datasets are used, covering various medical imaging modalities.

  • �� Model Training: WRN-16-8 models are trained on both the original and nonrobust feature datasets.

  • �� Adversarial Training: The TRADES loss function is used to enhance model robustness through adversarial training.

  • �� Experimental Design: Extensive hyperparameter search ensures model optimization.

  • �� Performance Evaluation: Models are evaluated on in-distribution and out-of-distribution test sets, using balanced accuracy as the primary metric.

Experiments

The experimental design includes the following aspects:


  • �� Datasets: Five MedMNIST datasets are selected, covering CT, radiography, ultrasound, and histopathology images.

  • �� Baseline Models: WRN-16-8 models are used as baselines, undergoing standard and adversarial training.

  • �� Evaluation Metrics: Balanced accuracy and AUC are used as primary evaluation metrics.

  • �� Hyperparameter Search: Extensive hyperparameter search ensures model optimization.

  • �� Adversarial Attacks: AutoAttack is used to evaluate the adversarial robustness of the models.

Results

The experimental results show:


  • �� In in-distribution tests, nonrobust feature models perform better than random guessing, demonstrating their predictive value. For instance, the nonrobust model achieves a balanced accuracy of 62% on PathMNIST.

  • �� Adversarially trained models perform excellently under controlled distribution shifts (MedMNIST-C), with balanced accuracy increasing from 78% to 89% on PathMNIST.

  • �� Nonrobust feature models perform poorly under adversarial attacks, with accuracy dropping to 4%, whereas adversarially trained models perform significantly better, with accuracy ranging from 57% to 73%.

Applications

The study's findings have significant implications in the following application scenarios:


  • �� Medical Imaging Analysis: Enhancing model robustness in out-of-distribution tests, improving model reliability and interpretability.

  • �� Clinical Decision Support: Applying robust features in medical imaging analysis to improve model performance in real-world applications.

  • �� Model Deployment: Balancing robustness and accuracy according to different application scenarios, optimizing model deployment strategies.

Limitations & Outlook

Despite the important findings, the study has the following limitations:


  • �� The study primarily focuses on two-dimensional low-resolution images, not involving three-dimensional or high-resolution imaging, which may limit the generalizability of the conclusions.

  • �� The evaluation does not cover natural distribution shifts; future studies should consider independently sourced datasets.

  • �� The experiments are limited to CNNs, and the performance of modern architectures like vision transformers has not been verified.

Plain Language Accessible to non-experts

Imagine you're in a kitchen, cooking a meal. The kitchen is full of tools and ingredients, some of which are easy to use, like knives and pans, while others require skill, like a fancy blender. In this analogy, the kitchen is like a deep learning model, and the tools and ingredients are the features the model learns.

In this kitchen, some tools, although not often used, are very useful in certain situations, like a can opener. These tools are like nonrobust features; they can help you quickly complete a task in some cases, but they may not be reliable in others.

However, if you rely only on these rarely used tools, you might run into trouble in some situations. For example, when you need to chop vegetables quickly, a can opener won't be much help. This is similar to medical imaging analysis, where nonrobust features, while performing well in some tests, may not be stable when facing changes.

Therefore, in the kitchen, you need to choose the right tools based on different needs, just like in model development, you need to balance robustness and accuracy, choosing the right features to improve the model's performance.

ELI14 Explained like you're 14

Hey there! Let's talk about a super cool study about medical images and artificial intelligence. Imagine you're playing a super complex game with lots of levels, each with different challenges. To win the game, you need to find some special tricks and tools.

In this study, scientists are like gamers, researching a technology called 'deep learning.' This tech is like a super weapon in the game, helping doctors analyze medical images like X-rays and CT scans.

But sometimes, these super weapons rely on 'nonrobust features,' like hidden skills in the game. These skills are useful in some levels but might get you into trouble in others.

So, scientists are working on making these super weapons more reliable, not just performing well in easy levels but also staying stable in tough ones. It's like finding a strategy in the game that can defeat both small monsters and big bosses!

Glossary

Deep Neural Network

A computational model mimicking the structure of the human brain, consisting of multiple layers, each with numerous neurons, used to process and learn complex patterns in data.

Used in this paper for medical image classification tasks.

Nonrobust Features

Input patterns highly sensitive to small adversarial perturbations and difficult to interpret, performing well in in-distribution tests but potentially degrading performance in out-of-distribution tests.

The core subject of study, exploring their role in medical imaging.

Adversarial Training

A method to improve model robustness by incorporating adversarial examples during training, aiming to enhance the model's resistance to adversarial attacks.

Used to improve model performance in out-of-distribution tests.

MedMNIST

A collection of datasets containing various medical imaging modalities, used to evaluate model performance in medical image classification tasks.

Used in this paper to experiment and verify the existence of nonrobust features.

Out-of-Distribution Testing

Evaluating model performance on data outside the training data distribution, aiming to test the model's generalization ability and robustness.

Used to verify the advantage of robust features in out-of-distribution tests.

Balanced Accuracy

An evaluation metric used to address class imbalance, calculated as the average recall rate of each class.

Used as the primary evaluation metric to measure model performance across different datasets.

Adversarial Attack

A technique to deceive models by making small perturbations to input data, causing incorrect predictions.

Used to test the robustness of models, especially the performance of nonrobust feature models.

WRN-16-8 (WideResNet-16-8)

An improved convolutional neural network architecture with a wider network structure to enhance model performance and stability.

Used as the baseline model for training and evaluation.

TRADES Loss Function

A loss function used in adversarial training to balance model performance on clean and adversarial data.

Used to enhance model robustness.

AutoAttack

A powerful adversarial attack method combining multiple parameter-free attack strategies to evaluate model adversarial robustness.

Used to verify the robustness of adversarially trained models.

Open Questions Unanswered questions from this research

  • 1 Although this paper reveals the widespread presence of nonrobust features in medical imaging, their role in three-dimensional or high-resolution imaging has not been fully studied. Future research should explore the performance of these features in more complex imaging.
  • 2 The current study focuses mainly on artificial distribution shifts, lacking evaluation of natural distribution shifts. Future research should consider using independently sourced datasets to verify model robustness in real-world scenarios.
  • 3 While adversarial training improves model robustness, it is computationally expensive. Future research should explore more efficient training methods to reduce computational resource consumption.
  • 4 This paper only uses CNNs for experiments, and the performance of modern architectures like vision transformers has not been verified. Future research should explore the potential of these emerging architectures in medical imaging analysis.
  • 5 The study focuses mainly on classification tasks; future research should extend to segmentation or detection tasks to verify the generalizability of these findings.
  • 6 The trade-off between robustness and accuracy requires further study, especially in selecting the best strategy for different application scenarios.
  • 7 The role of nonrobust features in improving standard test accuracy deserves further exploration, especially regarding performance differences across different datasets and tasks.

Applications

Immediate Applications

Medical Imaging Analysis

Enhancing model robustness in out-of-distribution tests, improving model reliability and interpretability. Applicable in fields like radiology and pathology.

Clinical Decision Support

Applying robust features in medical imaging analysis to improve model performance in real-world applications, aiding doctors in making more accurate diagnoses.

Model Deployment

Balancing robustness and accuracy according to different application scenarios, optimizing model deployment strategies, suitable for hospitals and research institutions.

Long-term Vision

Intelligent Healthcare Systems

Developing smarter, more reliable medical imaging analysis systems to support automated diagnosis and treatment recommendations, driving the digital transformation of the healthcare industry.

Personalized Medicine

Providing personalized diagnosis and treatment plans by analyzing patients' medical imaging data, improving the quality and efficiency of healthcare services.

Abstract

We study whether deep networks for medical imaging learn useful nonrobust features - predictive input patterns that are not human interpretable and highly susceptible to small adversarial perturbations - and how these features impact test performance. We show that models trained only on nonrobust features achieve well above chance accuracy across five MedMNIST classification tasks, confirming their predictive value in-distribution. Conversely, adversarially trained models that primarily rely on robust features sacrifice in-distribution accuracy but yield markedly better performance under controlled distribution shifts (MedMNIST-C). Overall, nonrobust features boost standard accuracy yet degrade out-of-distribution performance, revealing a practical robustness-accuracy trade-off in medical imaging classification tasks that should be tailored to the requirements of the deployment setting.

eess.IV cs.CV cs.LG

References (13)

Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks

Francesco Croce, Matthias Hein

2020 2323 citations View Analysis →

Robustness May Be at Odds with Accuracy

Dimitris Tsipras, Shibani Santurkar, Logan Engstrom et al.

2018 1931 citations View Analysis →

Wide Residual Networks

Sergey Zagoruyko, N. Komodakis

2016 8802 citations View Analysis →

Adversarial Examples Are Not Bugs, They Are Features

Andrew Ilyas, Shibani Santurkar, Dimitris Tsipras et al.

2019 2081 citations View Analysis →

RobustBench: a standardized adversarial robustness benchmark

Francesco Croce, Maksym Andriushchenko, Vikash Sehwag et al.

2020 877 citations View Analysis →

Overfitting in adversarially robust deep learning

Eric Wong, L. Rice, Zico Kolter

2020 926 citations View Analysis →

MedMNIST-C: Comprehensive benchmark and improved classifier robustness by simulating realistic image corruptions

Francesco Di Salvo, Sebastian Doerrich, Christian Ledig

2024 18 citations View Analysis →

Theoretically Principled Trade-off between Robustness and Accuracy

Hongyang Zhang, Yaodong Yu, Jiantao Jiao et al.

2019 2991 citations View Analysis →

Towards Deep Learning Models Resistant to Adversarial Attacks

A. Ma̧dry, Aleksandar Makelov, Ludwig Schmidt et al.

2017 14493 citations View Analysis →

Better Diffusion Models Further Improve Adversarial Training

Zekai Wang, Tianyu Pang, Chao Du et al.

2023 299 citations View Analysis →

MedMNIST v2 - A large-scale lightweight benchmark for 2D and 3D biomedical image classification

Jiancheng Yang, Rui Shi, D. Wei et al.

2021 1202 citations View Analysis →

What Can the Neural Tangent Kernel Tell Us About Adversarial Robustness?

Nikolaos Tsilivis, J. Kempe

2022 27 citations View Analysis →

Bag of Tricks for Adversarial Training

Tianyu Pang, Xiao Yang, Yinpeng Dong et al.

2020 297 citations View Analysis →