Statistical diagnosis system for adversarial examples

DeepNeuralNetworks (DNNs) are powerful to the classification tasks, finding the potential links between dataset with high accuracy and speed. However, the DNNs are also fragile to intentionally produced adversarial attacks, especially in the field of image analysis where also the concept of adver...

Full description

Bibliographic Details
Main Author: Wu, Yuting
Other Authors: Wang Dan Wei
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2020
Subjects:
Online Access:https://hdl.handle.net/10356/140900
_version_ 1811682270672584704
author Wu, Yuting
author2 Wang Dan Wei
author_facet Wang Dan Wei
Wu, Yuting
author_sort Wu, Yuting
collection NTU
description DeepNeuralNetworks (DNNs) are powerful to the classification tasks, finding the potential links between dataset with high accuracy and speed. However, the DNNs are also fragile to intentionally produced adversarial attacks, especially in the field of image analysis where also the concept of adversarial examples first emerged. These adversarial perturbations are designed to be quasi-imperceptible to human vision but can easily fool the deep models with high confidence. This situation aroused researchers’ great interest in detection and defense of adversarial examples to improve the reliability of the deep neural networks which would play an important role in safety and security systems in the coming future. In the view of that, this work will first give a brief view of common attacks on Mnist and Cifar-10 dataset with a general concept on what adversarial examples are and how to generate it. After that, different kinds of defense methods will be introduced, and we will mainly focus on statistical defense way. Experiments are conducted to evaluate the performance of those existing defense methods on their merits and demerits. In chapter 4, an improvement on Principal Component Analysis with Gaussian Mixture Model method is proposed which enables it to detect adversarial examples from dataset attacked by C&W attack. In chapter 5, this dissertation proposes an improvedKernel-Density-Estimation detection method based on a Deep Graph infomax. We assume that a simple modification on loss function, with an extra loss that maximizes the mutual information between images and their deep representations, the DNN models could extract more key information from input images to their deep feature map. After the modification, it would help the model detect the unique feature of adversarial examples and improve the detection result. The experiment in chapter 5 has demonstrated the verification of our assumption.
first_indexed 2024-10-01T03:54:11Z
format Thesis-Master by Coursework
id ntu-10356/140900
institution Nanyang Technological University
language English
last_indexed 2024-10-01T03:54:11Z
publishDate 2020
publisher Nanyang Technological University
record_format dspace
spelling ntu-10356/1409002023-07-04T16:19:44Z Statistical diagnosis system for adversarial examples Wu, Yuting Wang Dan Wei School of Electrical and Electronic Engineering EDWWANG@ntu.edu.sg Engineering::Electrical and electronic engineering DeepNeuralNetworks (DNNs) are powerful to the classification tasks, finding the potential links between dataset with high accuracy and speed. However, the DNNs are also fragile to intentionally produced adversarial attacks, especially in the field of image analysis where also the concept of adversarial examples first emerged. These adversarial perturbations are designed to be quasi-imperceptible to human vision but can easily fool the deep models with high confidence. This situation aroused researchers’ great interest in detection and defense of adversarial examples to improve the reliability of the deep neural networks which would play an important role in safety and security systems in the coming future. In the view of that, this work will first give a brief view of common attacks on Mnist and Cifar-10 dataset with a general concept on what adversarial examples are and how to generate it. After that, different kinds of defense methods will be introduced, and we will mainly focus on statistical defense way. Experiments are conducted to evaluate the performance of those existing defense methods on their merits and demerits. In chapter 4, an improvement on Principal Component Analysis with Gaussian Mixture Model method is proposed which enables it to detect adversarial examples from dataset attacked by C&W attack. In chapter 5, this dissertation proposes an improvedKernel-Density-Estimation detection method based on a Deep Graph infomax. We assume that a simple modification on loss function, with an extra loss that maximizes the mutual information between images and their deep representations, the DNN models could extract more key information from input images to their deep feature map. After the modification, it would help the model detect the unique feature of adversarial examples and improve the detection result. The experiment in chapter 5 has demonstrated the verification of our assumption. Master of Science (Computer Control and Automation) 2020-06-02T12:45:45Z 2020-06-02T12:45:45Z 2020 Thesis-Master by Coursework https://hdl.handle.net/10356/140900 en application/pdf Nanyang Technological University
spellingShingle Engineering::Electrical and electronic engineering
Wu, Yuting
Statistical diagnosis system for adversarial examples
title Statistical diagnosis system for adversarial examples
title_full Statistical diagnosis system for adversarial examples
title_fullStr Statistical diagnosis system for adversarial examples
title_full_unstemmed Statistical diagnosis system for adversarial examples
title_short Statistical diagnosis system for adversarial examples
title_sort statistical diagnosis system for adversarial examples
topic Engineering::Electrical and electronic engineering
url https://hdl.handle.net/10356/140900
work_keys_str_mv AT wuyuting statisticaldiagnosissystemforadversarialexamples