Out of Distribution Detection and Adversarial Attacks on Deep Neural Networks for Robust Medical Image Analysis
Abstract
Deep learning models have become a popular choice for medical image analysis. However, the poor generalization performance of deep learning models limits them from being deployed in the real world as robustness is critical for medical applications. For instance, the state-of-the-art Convolutional Neural Networks (CNNs) fail to detect samples drawn statistically far away from the training distribution or adversarially. In this work, we experimentally evaluate the robustness of a Mahalanobis distance-based confidence score, a simple yet effective method for detecting abnormal input samples, in classifying malaria parasitized cells and uninfected cells. Results indicated that the Mahalanobis confidence score detector exhibits improved performance and robustness of deep learning models, and achieves state-of-the-art performance on both out-of-distribution and adversarial samples.
Cite
Text
Uwimana and Senanayake. "Out of Distribution Detection and Adversarial Attacks on Deep Neural Networks for Robust Medical Image Analysis." ICML 2021 Workshops: AML, 2021.Markdown
[Uwimana and Senanayake. "Out of Distribution Detection and Adversarial Attacks on Deep Neural Networks for Robust Medical Image Analysis." ICML 2021 Workshops: AML, 2021.](https://mlanthology.org/icmlw/2021/uwimana2021icmlw-out/)BibTeX
@inproceedings{uwimana2021icmlw-out,
title = {{Out of Distribution Detection and Adversarial Attacks on Deep Neural Networks for Robust Medical Image Analysis}},
author = {Uwimana, Anisie and Senanayake, Ransalu},
booktitle = {ICML 2021 Workshops: AML},
year = {2021},
url = {https://mlanthology.org/icmlw/2021/uwimana2021icmlw-out/}
}