TY - JOUR
T1 - Effects of Image Processing Operations on Adversarial Noise and Their Use in Detecting and Correcting Adversarial Images
AU - Nguyen, Huy H.
AU - Kuribayashi, Minoru
AU - Yamagishi, Junichi
AU - Echizen, Isao
N1 - Funding Information:
This research was supported by JSPS KAKENHI Grants JP16H06302, JP17H04687, JP18H04120, JP18H04112, JP18KT0051, and JP19K22846 and by JST CREST Grants JPMJCR18A6 and JPMJCR20D3, Japan.
Publisher Copyright:
Copyright © 2022 The Institute of Electronics, Information and Communication Engineers.
PY - 2022
Y1 - 2022
N2 - Deep neural networks (DNNs) have achieved excellent performance on several tasks and have been widely applied in both academia and industry. However, DNNs are vulnerable to adversarial machine learning attacks in which noise is added to the input to change the networks' output. Consequently, DNN-based mission-critical applications such as those used in self-driving vehicles have reduced reliability and could cause severe accidents and damage. Moreover, adversarial examples could be used to poison DNN training data, resulting in corruptions of trained models. Besides the need for detecting adversarial examples, correcting them is important for restoring data and system functionality to normal. We have developed methods for detecting and correcting adversarial images that use multiple image processing operations with multiple parameter values. For detection, we devised a statistical-based method that outperforms the feature squeezing method. For correction, we devised a method that uses for the first time two levels of correction. The first level is label correction, with the focus on restoring the adversarial images' original predicted labels (for use in the current task). The second level is image correction, with the focus on both the correctness and quality of the corrected images (for use in the current and other tasks). Our experiments demonstrated that the correction method could correct nearly 90% of the adversarial images created by classical adversarial attacks and affected only about 2% of the normal images.
AB - Deep neural networks (DNNs) have achieved excellent performance on several tasks and have been widely applied in both academia and industry. However, DNNs are vulnerable to adversarial machine learning attacks in which noise is added to the input to change the networks' output. Consequently, DNN-based mission-critical applications such as those used in self-driving vehicles have reduced reliability and could cause severe accidents and damage. Moreover, adversarial examples could be used to poison DNN training data, resulting in corruptions of trained models. Besides the need for detecting adversarial examples, correcting them is important for restoring data and system functionality to normal. We have developed methods for detecting and correcting adversarial images that use multiple image processing operations with multiple parameter values. For detection, we devised a statistical-based method that outperforms the feature squeezing method. For correction, we devised a method that uses for the first time two levels of correction. The first level is label correction, with the focus on restoring the adversarial images' original predicted labels (for use in the current task). The second level is image correction, with the focus on both the correctness and quality of the corrected images (for use in the current and other tasks). Our experiments demonstrated that the correction method could correct nearly 90% of the adversarial images created by classical adversarial attacks and affected only about 2% of the normal images.
KW - Adversarial machine learning
KW - Correcting adversarial image
KW - Data cleansing
KW - Deep neural network
KW - Detecting adversarial image
KW - Image processing operation
UR - http://www.scopus.com/inward/record.url?scp=85123417485&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85123417485&partnerID=8YFLogxK
U2 - 10.1587/transinf.2021MUP0005
DO - 10.1587/transinf.2021MUP0005
M3 - Article
AN - SCOPUS:85123417485
SN - 0916-8532
VL - E105D
SP - 65
EP - 77
JO - IEICE Transactions on Information and Systems
JF - IEICE Transactions on Information and Systems
IS - 1
ER -