Medical imaging deep learning with differential privacy

Abstract The successful training of deep learning models for diagnostic deployment in medical imaging applications requires large volumes of data. Such data cannot be procured without consideration for patient privacy, mandated both by legal regulations and ethical requirements of the medical profes...

Full description

Bibliographic Details
Main Authors: Alexander Ziller, Dmitrii Usynin, Rickmer Braren, Marcus Makowski, Daniel Rueckert, Georgios Kaissis
Format: Article
Language:English
Published: Nature Publishing Group 2021-06-01
Series:Scientific Reports
Online Access:https://doi.org/10.1038/s41598-021-93030-0
id doaj-3af80b8b1cd44f2eab82247961c4d757
record_format Article
spelling doaj-3af80b8b1cd44f2eab82247961c4d7572021-07-04T11:26:29ZengNature Publishing GroupScientific Reports2045-23222021-06-011111810.1038/s41598-021-93030-0Medical imaging deep learning with differential privacyAlexander Ziller0Dmitrii Usynin1Rickmer Braren2Marcus Makowski3Daniel Rueckert4Georgios Kaissis5Institute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, School of Medicine, Technical University of MunichInstitute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, School of Medicine, Technical University of MunichInstitute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, School of Medicine, Technical University of MunichInstitute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, School of Medicine, Technical University of MunichArtificial Intelligence in Medicine and Healthcare, Technical University of MunichInstitute for Diagnostic and Interventional Radiology, Klinikum rechts der Isar, School of Medicine, Technical University of MunichAbstract The successful training of deep learning models for diagnostic deployment in medical imaging applications requires large volumes of data. Such data cannot be procured without consideration for patient privacy, mandated both by legal regulations and ethical requirements of the medical profession. Differential privacy (DP) enables the provision of information-theoretic privacy guarantees to patients and can be implemented in the setting of deep neural network training through the differentially private stochastic gradient descent (DP-SGD) algorithm. We here present deepee, a free-and-open-source framework for differentially private deep learning for use with the PyTorch deep learning framework. Our framework is based on parallelised execution of neural network operations to obtain and modify the per-sample gradients. The process is efficiently abstracted via a data structure maintaining shared memory references to neural network weights to maintain memory efficiency. We furthermore offer specialised data loading procedures and privacy budget accounting based on the Gaussian Differential Privacy framework, as well as automated modification of the user-supplied neural network architectures to ensure DP-conformity of its layers. We benchmark our framework’s computational performance against other open-source DP frameworks and evaluate its application on the paediatric pneumonia dataset, an image classification task and on the Medical Segmentation Decathlon Liver dataset in the task of medical image segmentation. We find that neural network training with rigorous privacy guarantees is possible while maintaining acceptable classification performance and excellent segmentation performance. Our framework compares favourably to related work with respect to memory consumption and computational performance. Our work presents an open-source software framework for differentially private deep learning, which we demonstrate in medical imaging analysis tasks. It serves to further the utilisation of privacy-enhancing techniques in medicine and beyond in order to assist researchers and practitioners in addressing the numerous outstanding challenges towards their widespread implementation.https://doi.org/10.1038/s41598-021-93030-0
collection DOAJ
language English
format Article
sources DOAJ
author Alexander Ziller
Dmitrii Usynin
Rickmer Braren
Marcus Makowski
Daniel Rueckert
Georgios Kaissis
spellingShingle Alexander Ziller
Dmitrii Usynin
Rickmer Braren
Marcus Makowski
Daniel Rueckert
Georgios Kaissis
Medical imaging deep learning with differential privacy
Scientific Reports
author_facet Alexander Ziller
Dmitrii Usynin
Rickmer Braren
Marcus Makowski
Daniel Rueckert
Georgios Kaissis
author_sort Alexander Ziller
title Medical imaging deep learning with differential privacy
title_short Medical imaging deep learning with differential privacy
title_full Medical imaging deep learning with differential privacy
title_fullStr Medical imaging deep learning with differential privacy
title_full_unstemmed Medical imaging deep learning with differential privacy
title_sort medical imaging deep learning with differential privacy
publisher Nature Publishing Group
series Scientific Reports
issn 2045-2322
publishDate 2021-06-01
description Abstract The successful training of deep learning models for diagnostic deployment in medical imaging applications requires large volumes of data. Such data cannot be procured without consideration for patient privacy, mandated both by legal regulations and ethical requirements of the medical profession. Differential privacy (DP) enables the provision of information-theoretic privacy guarantees to patients and can be implemented in the setting of deep neural network training through the differentially private stochastic gradient descent (DP-SGD) algorithm. We here present deepee, a free-and-open-source framework for differentially private deep learning for use with the PyTorch deep learning framework. Our framework is based on parallelised execution of neural network operations to obtain and modify the per-sample gradients. The process is efficiently abstracted via a data structure maintaining shared memory references to neural network weights to maintain memory efficiency. We furthermore offer specialised data loading procedures and privacy budget accounting based on the Gaussian Differential Privacy framework, as well as automated modification of the user-supplied neural network architectures to ensure DP-conformity of its layers. We benchmark our framework’s computational performance against other open-source DP frameworks and evaluate its application on the paediatric pneumonia dataset, an image classification task and on the Medical Segmentation Decathlon Liver dataset in the task of medical image segmentation. We find that neural network training with rigorous privacy guarantees is possible while maintaining acceptable classification performance and excellent segmentation performance. Our framework compares favourably to related work with respect to memory consumption and computational performance. Our work presents an open-source software framework for differentially private deep learning, which we demonstrate in medical imaging analysis tasks. It serves to further the utilisation of privacy-enhancing techniques in medicine and beyond in order to assist researchers and practitioners in addressing the numerous outstanding challenges towards their widespread implementation.
url https://doi.org/10.1038/s41598-021-93030-0
work_keys_str_mv AT alexanderziller medicalimagingdeeplearningwithdifferentialprivacy
AT dmitriiusynin medicalimagingdeeplearningwithdifferentialprivacy
AT rickmerbraren medicalimagingdeeplearningwithdifferentialprivacy
AT marcusmakowski medicalimagingdeeplearningwithdifferentialprivacy
AT danielrueckert medicalimagingdeeplearningwithdifferentialprivacy
AT georgioskaissis medicalimagingdeeplearningwithdifferentialprivacy
_version_ 1721320285992386560