Detection of attacks on artificial intelligence systems

Artificial intelligence (AI) is gradually and profoundly changing production and life, generally used in various fields such as visual information processing, autonomous systems, safety diagnosis and protection. Security issues will eventually become the biggest challenge. The adversarial attack is...

Full description

Bibliographic Details
Main Author: Pan, Siyu
Other Authors: Wen Bihan
Format: Thesis-Master by Coursework
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/152977
_version_ 1826114725861130240
author Pan, Siyu
author2 Wen Bihan
author_facet Wen Bihan
Pan, Siyu
author_sort Pan, Siyu
collection NTU
description Artificial intelligence (AI) is gradually and profoundly changing production and life, generally used in various fields such as visual information processing, autonomous systems, safety diagnosis and protection. Security issues will eventually become the biggest challenge. The adversarial attack is a powerful security threat to Deep Neural Networks (DNNs). This dissertation focuses on a passive defence method -- the detection of adversarial samples. The adversarial sample is essentially different from the normal sample. The dimension of the high-dimensional continuous space in which it is located is much larger than the intrinsic dimensionality of any given data submanifold. Focusing on the Local Intrinsic Dimensionality (LID), a better detector -- LID-based classifier is studied. Four attack methods were used to conduct experiments on two common datasets. The experiments show that the LID-based classifier has a great performance improvement than the single-characteristic classifier based on Kernel Density (KD) and Bayesian Network Uncertainty (BNU) and the combined classifier based on KD&BNU. The improvement is up to 37.44% for the single-characteristic classifiers, and up to 18.65% for the combined classifiers. Then it proves that the LID-based classifier trained based on one attack can detect adversarial samples generated by other attack methods. The classifiers trained on weaker attacks will perform better in the face of adversarial samples generated by other stronger attacks than when testing under the same attack, and vice versa. It fully proves that the LID-based classifier is a very effective means to detect adversarial samples, with certain universality and transferability. Finally, the dissertation also makes a prediction and recommendation for the possible direction of the future work.
first_indexed 2024-10-01T03:44:10Z
format Thesis-Master by Coursework
id ntu-10356/152977
institution Nanyang Technological University
language English
last_indexed 2024-10-01T03:44:10Z
publishDate 2021
publisher Nanyang Technological University
record_format dspace
spelling ntu-10356/1529772023-07-04T16:37:07Z Detection of attacks on artificial intelligence systems Pan, Siyu Wen Bihan School of Electrical and Electronic Engineering bihan.wen@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Artificial intelligence (AI) is gradually and profoundly changing production and life, generally used in various fields such as visual information processing, autonomous systems, safety diagnosis and protection. Security issues will eventually become the biggest challenge. The adversarial attack is a powerful security threat to Deep Neural Networks (DNNs). This dissertation focuses on a passive defence method -- the detection of adversarial samples. The adversarial sample is essentially different from the normal sample. The dimension of the high-dimensional continuous space in which it is located is much larger than the intrinsic dimensionality of any given data submanifold. Focusing on the Local Intrinsic Dimensionality (LID), a better detector -- LID-based classifier is studied. Four attack methods were used to conduct experiments on two common datasets. The experiments show that the LID-based classifier has a great performance improvement than the single-characteristic classifier based on Kernel Density (KD) and Bayesian Network Uncertainty (BNU) and the combined classifier based on KD&BNU. The improvement is up to 37.44% for the single-characteristic classifiers, and up to 18.65% for the combined classifiers. Then it proves that the LID-based classifier trained based on one attack can detect adversarial samples generated by other attack methods. The classifiers trained on weaker attacks will perform better in the face of adversarial samples generated by other stronger attacks than when testing under the same attack, and vice versa. It fully proves that the LID-based classifier is a very effective means to detect adversarial samples, with certain universality and transferability. Finally, the dissertation also makes a prediction and recommendation for the possible direction of the future work. Master of Science (Signal Processing) 2021-10-26T04:30:55Z 2021-10-26T04:30:55Z 2021 Thesis-Master by Coursework Pan, S. (2021). Detection of attacks on artificial intelligence systems. Master's thesis, Nanyang Technological University, Singapore. https://hdl.handle.net/10356/152977 https://hdl.handle.net/10356/152977 en application/pdf Nanyang Technological University
spellingShingle Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
Pan, Siyu
Detection of attacks on artificial intelligence systems
title Detection of attacks on artificial intelligence systems
title_full Detection of attacks on artificial intelligence systems
title_fullStr Detection of attacks on artificial intelligence systems
title_full_unstemmed Detection of attacks on artificial intelligence systems
title_short Detection of attacks on artificial intelligence systems
title_sort detection of attacks on artificial intelligence systems
topic Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence
url https://hdl.handle.net/10356/152977
work_keys_str_mv AT pansiyu detectionofattacksonartificialintelligencesystems