Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW

This study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and anal...

Full description

Bibliographic Details
Main Authors: William Villegas-Ch, Angel Jaramillo-Alcázar, Sergio Luján-Mora
Format: Article
Language:English
Published: MDPI AG 2024-01-01
Series:Big Data and Cognitive Computing
Subjects:
Online Access:https://www.mdpi.com/2504-2289/8/1/8
_version_ 1797344625556455424
author William Villegas-Ch
Angel Jaramillo-Alcázar
Sergio Luján-Mora
author_facet William Villegas-Ch
Angel Jaramillo-Alcázar
Sergio Luján-Mora
author_sort William Villegas-Ch
collection DOAJ
description This study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and analyze their impact on the model’s classification accuracy. Additionally, image manipulation techniques were investigated as defensive measures against adversarial attacks. The results highlighted the model’s vulnerability to conflicting examples: the Fast Gradient Signed Method effectively altered the original classifications, while the Carlini and Wagner method proved less effective. Promising approaches such as noise reduction, image compression, and Gaussian blurring were presented as effective countermeasures. These findings underscore the importance of addressing the vulnerability of machine learning models and the need to develop robust defenses against adversarial examples. This article emphasizes the urgency of addressing the threat posed by harmful standards in machine learning models, highlighting the relevance of implementing effective countermeasures and image manipulation techniques to mitigate the effects of adversarial attacks. These efforts are crucial to safeguarding model integrity and trust in an environment marked by constantly evolving hostile threats. An average 25% decrease in accuracy was observed for the VGG16 model when exposed to the Fast Gradient Signed Method and Projected Gradient Descent attacks, and an even more significant 35% decrease with the Carlini and Wagner method.
first_indexed 2024-03-08T11:05:26Z
format Article
id doaj.art-88b05e1082b747319eb14c77421915ef
institution Directory Open Access Journal
issn 2504-2289
language English
last_indexed 2024-03-08T11:05:26Z
publishDate 2024-01-01
publisher MDPI AG
record_format Article
series Big Data and Cognitive Computing
spelling doaj.art-88b05e1082b747319eb14c77421915ef2024-01-26T15:05:33ZengMDPI AGBig Data and Cognitive Computing2504-22892024-01-0181810.3390/bdcc8010008Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CWWilliam Villegas-Ch0Angel Jaramillo-Alcázar1Sergio Luján-Mora2Escuela de Ingeniería en Ciberseguridad, Facultad de Ingenierías Ciencias Aplicadas, Universidad de Las Américas, Quito 170125, EcuadorEscuela de Ingeniería en Ciberseguridad, Facultad de Ingenierías Ciencias Aplicadas, Universidad de Las Américas, Quito 170125, EcuadorDepartamento de Lenguajes y Sistemas Informáticos, Universidad de Alicante, 03690 Alicante, SpainThis study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and analyze their impact on the model’s classification accuracy. Additionally, image manipulation techniques were investigated as defensive measures against adversarial attacks. The results highlighted the model’s vulnerability to conflicting examples: the Fast Gradient Signed Method effectively altered the original classifications, while the Carlini and Wagner method proved less effective. Promising approaches such as noise reduction, image compression, and Gaussian blurring were presented as effective countermeasures. These findings underscore the importance of addressing the vulnerability of machine learning models and the need to develop robust defenses against adversarial examples. This article emphasizes the urgency of addressing the threat posed by harmful standards in machine learning models, highlighting the relevance of implementing effective countermeasures and image manipulation techniques to mitigate the effects of adversarial attacks. These efforts are crucial to safeguarding model integrity and trust in an environment marked by constantly evolving hostile threats. An average 25% decrease in accuracy was observed for the VGG16 model when exposed to the Fast Gradient Signed Method and Projected Gradient Descent attacks, and an even more significant 35% decrease with the Carlini and Wagner method.https://www.mdpi.com/2504-2289/8/1/8adversary examplesrobustness of modelscountermeasures
spellingShingle William Villegas-Ch
Angel Jaramillo-Alcázar
Sergio Luján-Mora
Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
Big Data and Cognitive Computing
adversary examples
robustness of models
countermeasures
title Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
title_full Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
title_fullStr Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
title_full_unstemmed Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
title_short Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
title_sort evaluating the robustness of deep learning models against adversarial attacks an analysis with fgsm pgd and cw
topic adversary examples
robustness of models
countermeasures
url https://www.mdpi.com/2504-2289/8/1/8
work_keys_str_mv AT williamvillegasch evaluatingtherobustnessofdeeplearningmodelsagainstadversarialattacksananalysiswithfgsmpgdandcw
AT angeljaramilloalcazar evaluatingtherobustnessofdeeplearningmodelsagainstadversarialattacksananalysiswithfgsmpgdandcw
AT sergiolujanmora evaluatingtherobustnessofdeeplearningmodelsagainstadversarialattacksananalysiswithfgsmpgdandcw