Black-Box Evasion Attack Method Based on Confidence Score of Benign Samples

Recently, malware detection models based on deep learning have gradually replaced manual analysis as the first line of defense for anti-malware systems. However, it has been shown that these models are vulnerable to a specific class of inputs called adversarial examples. It is possible to evade the...

Full description

Bibliographic Details
Main Authors: Shaohan Wu, Jingfeng Xue, Yong Wang, Zixiao Kong
Format: Article
Language:English
Published: MDPI AG 2023-05-01
Series:Electronics
Subjects:
Online Access:https://www.mdpi.com/2079-9292/12/11/2346
_version_ 1797597725372448768
author Shaohan Wu
Jingfeng Xue
Yong Wang
Zixiao Kong
author_facet Shaohan Wu
Jingfeng Xue
Yong Wang
Zixiao Kong
author_sort Shaohan Wu
collection DOAJ
description Recently, malware detection models based on deep learning have gradually replaced manual analysis as the first line of defense for anti-malware systems. However, it has been shown that these models are vulnerable to a specific class of inputs called adversarial examples. It is possible to evade the detection model by adding some carefully crafted tiny perturbations to the malicious samples without changing the sample functions. Most of the adversarial example generation methods ignore the information contained in the detection results of benign samples from detection models. Our method extracts sequence fragments called benign payload from benign samples based on detection results and uses an RNN generative model to learn benign features embedded in these sequences. Then, we use the end of the original malicious sample as input to generate an adversarial perturbation that reduces the malicious probability of the sample and append it to the end of the sample to generate an adversarial sample. According to different adversarial scenarios, we propose two different generation strategies, which are the one-time generation method and the iterative generation method. Under different query times and append scale constraints, the maximum evasion success rate can reach 90.8%.
first_indexed 2024-03-11T03:09:33Z
format Article
id doaj.art-678976fe168f42418169c4a9138206d3
institution Directory Open Access Journal
issn 2079-9292
language English
last_indexed 2024-03-11T03:09:33Z
publishDate 2023-05-01
publisher MDPI AG
record_format Article
series Electronics
spelling doaj.art-678976fe168f42418169c4a9138206d32023-11-18T07:43:43ZengMDPI AGElectronics2079-92922023-05-011211234610.3390/electronics12112346Black-Box Evasion Attack Method Based on Confidence Score of Benign SamplesShaohan Wu0Jingfeng Xue1Yong Wang2Zixiao Kong3School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, ChinaSchool of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, ChinaSchool of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, ChinaSchool of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, ChinaRecently, malware detection models based on deep learning have gradually replaced manual analysis as the first line of defense for anti-malware systems. However, it has been shown that these models are vulnerable to a specific class of inputs called adversarial examples. It is possible to evade the detection model by adding some carefully crafted tiny perturbations to the malicious samples without changing the sample functions. Most of the adversarial example generation methods ignore the information contained in the detection results of benign samples from detection models. Our method extracts sequence fragments called benign payload from benign samples based on detection results and uses an RNN generative model to learn benign features embedded in these sequences. Then, we use the end of the original malicious sample as input to generate an adversarial perturbation that reduces the malicious probability of the sample and append it to the end of the sample to generate an adversarial sample. According to different adversarial scenarios, we propose two different generation strategies, which are the one-time generation method and the iterative generation method. Under different query times and append scale constraints, the maximum evasion success rate can reach 90.8%.https://www.mdpi.com/2079-9292/12/11/2346adversarial examplesevasion attackmalware detectionartificial intelligence security
spellingShingle Shaohan Wu
Jingfeng Xue
Yong Wang
Zixiao Kong
Black-Box Evasion Attack Method Based on Confidence Score of Benign Samples
Electronics
adversarial examples
evasion attack
malware detection
artificial intelligence security
title Black-Box Evasion Attack Method Based on Confidence Score of Benign Samples
title_full Black-Box Evasion Attack Method Based on Confidence Score of Benign Samples
title_fullStr Black-Box Evasion Attack Method Based on Confidence Score of Benign Samples
title_full_unstemmed Black-Box Evasion Attack Method Based on Confidence Score of Benign Samples
title_short Black-Box Evasion Attack Method Based on Confidence Score of Benign Samples
title_sort black box evasion attack method based on confidence score of benign samples
topic adversarial examples
evasion attack
malware detection
artificial intelligence security
url https://www.mdpi.com/2079-9292/12/11/2346
work_keys_str_mv AT shaohanwu blackboxevasionattackmethodbasedonconfidencescoreofbenignsamples
AT jingfengxue blackboxevasionattackmethodbasedonconfidencescoreofbenignsamples
AT yongwang blackboxevasionattackmethodbasedonconfidencescoreofbenignsamples
AT zixiaokong blackboxevasionattackmethodbasedonconfidencescoreofbenignsamples