Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW
This study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and anal...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2024-01-01
|
Series: | Big Data and Cognitive Computing |
Subjects: | |
Online Access: | https://www.mdpi.com/2504-2289/8/1/8 |
_version_ | 1797344625556455424 |
---|---|
author | William Villegas-Ch Angel Jaramillo-Alcázar Sergio Luján-Mora |
author_facet | William Villegas-Ch Angel Jaramillo-Alcázar Sergio Luján-Mora |
author_sort | William Villegas-Ch |
collection | DOAJ |
description | This study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and analyze their impact on the model’s classification accuracy. Additionally, image manipulation techniques were investigated as defensive measures against adversarial attacks. The results highlighted the model’s vulnerability to conflicting examples: the Fast Gradient Signed Method effectively altered the original classifications, while the Carlini and Wagner method proved less effective. Promising approaches such as noise reduction, image compression, and Gaussian blurring were presented as effective countermeasures. These findings underscore the importance of addressing the vulnerability of machine learning models and the need to develop robust defenses against adversarial examples. This article emphasizes the urgency of addressing the threat posed by harmful standards in machine learning models, highlighting the relevance of implementing effective countermeasures and image manipulation techniques to mitigate the effects of adversarial attacks. These efforts are crucial to safeguarding model integrity and trust in an environment marked by constantly evolving hostile threats. An average 25% decrease in accuracy was observed for the VGG16 model when exposed to the Fast Gradient Signed Method and Projected Gradient Descent attacks, and an even more significant 35% decrease with the Carlini and Wagner method. |
first_indexed | 2024-03-08T11:05:26Z |
format | Article |
id | doaj.art-88b05e1082b747319eb14c77421915ef |
institution | Directory Open Access Journal |
issn | 2504-2289 |
language | English |
last_indexed | 2024-03-08T11:05:26Z |
publishDate | 2024-01-01 |
publisher | MDPI AG |
record_format | Article |
series | Big Data and Cognitive Computing |
spelling | doaj.art-88b05e1082b747319eb14c77421915ef2024-01-26T15:05:33ZengMDPI AGBig Data and Cognitive Computing2504-22892024-01-0181810.3390/bdcc8010008Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CWWilliam Villegas-Ch0Angel Jaramillo-Alcázar1Sergio Luján-Mora2Escuela de Ingeniería en Ciberseguridad, Facultad de Ingenierías Ciencias Aplicadas, Universidad de Las Américas, Quito 170125, EcuadorEscuela de Ingeniería en Ciberseguridad, Facultad de Ingenierías Ciencias Aplicadas, Universidad de Las Américas, Quito 170125, EcuadorDepartamento de Lenguajes y Sistemas Informáticos, Universidad de Alicante, 03690 Alicante, SpainThis study evaluated the generation of adversarial examples and the subsequent robustness of an image classification model. The attacks were performed using the Fast Gradient Sign method, the Projected Gradient Descent method, and the Carlini and Wagner attack to perturb the original images and analyze their impact on the model’s classification accuracy. Additionally, image manipulation techniques were investigated as defensive measures against adversarial attacks. The results highlighted the model’s vulnerability to conflicting examples: the Fast Gradient Signed Method effectively altered the original classifications, while the Carlini and Wagner method proved less effective. Promising approaches such as noise reduction, image compression, and Gaussian blurring were presented as effective countermeasures. These findings underscore the importance of addressing the vulnerability of machine learning models and the need to develop robust defenses against adversarial examples. This article emphasizes the urgency of addressing the threat posed by harmful standards in machine learning models, highlighting the relevance of implementing effective countermeasures and image manipulation techniques to mitigate the effects of adversarial attacks. These efforts are crucial to safeguarding model integrity and trust in an environment marked by constantly evolving hostile threats. An average 25% decrease in accuracy was observed for the VGG16 model when exposed to the Fast Gradient Signed Method and Projected Gradient Descent attacks, and an even more significant 35% decrease with the Carlini and Wagner method.https://www.mdpi.com/2504-2289/8/1/8adversary examplesrobustness of modelscountermeasures |
spellingShingle | William Villegas-Ch Angel Jaramillo-Alcázar Sergio Luján-Mora Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW Big Data and Cognitive Computing adversary examples robustness of models countermeasures |
title | Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW |
title_full | Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW |
title_fullStr | Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW |
title_full_unstemmed | Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW |
title_short | Evaluating the Robustness of Deep Learning Models against Adversarial Attacks: An Analysis with FGSM, PGD and CW |
title_sort | evaluating the robustness of deep learning models against adversarial attacks an analysis with fgsm pgd and cw |
topic | adversary examples robustness of models countermeasures |
url | https://www.mdpi.com/2504-2289/8/1/8 |
work_keys_str_mv | AT williamvillegasch evaluatingtherobustnessofdeeplearningmodelsagainstadversarialattacksananalysiswithfgsmpgdandcw AT angeljaramilloalcazar evaluatingtherobustnessofdeeplearningmodelsagainstadversarialattacksananalysiswithfgsmpgdandcw AT sergiolujanmora evaluatingtherobustnessofdeeplearningmodelsagainstadversarialattacksananalysiswithfgsmpgdandcw |