Generating adversarial images to monitor the training state of a CNN model

Deep neural networks have shown effectiveness in many applications, however, in regulated applications like automotive or medicine, quality guarantees are required. Thus, it is important to understand the robustness of the solutions to perturbations in the input space. In order to identify the vulne...

Full description

Bibliographic Details
Published in:Current Directions in Biomedical Engineering
Main Authors: Ding Ning, Möller Knut
Format: Article
Language:English
Published: De Gruyter 2021-10-01
Subjects:
Online Access:https://doi.org/10.1515/cdbme-2021-2077
Description
Summary:Deep neural networks have shown effectiveness in many applications, however, in regulated applications like automotive or medicine, quality guarantees are required. Thus, it is important to understand the robustness of the solutions to perturbations in the input space. In order to identify the vulnerability of a trained classification model and evaluate the effect of different perturbations in the input on the output class, two different methods to generate adversarial examples were implemented. The adversarial images created were developed into a robustness index to monitor the training state and safety of a convolutional neural network model. In the future work, some generated adversarial images will be included into the training phase to improve the model robustness.
ISSN:2364-5504