Selective real‐time adversarial perturbations against deep reinforcement learning agents

Abstract Recent work has shown that deep reinforcement learning (DRL) is vulnerable to adversarial attacks, so that exploiting vulnerabilities in DRL systems through adversarial attack techniques has become a necessary prerequisite for building robust DRL systems. Compared to traditional deep learni...

Full description

Bibliographic Details
Main Authors: Hongjin Yao, Yisheng Li, Yunpeng Sun, Zhichao Lian
Format: Article
Language:English
Published: Wiley 2024-03-01
Series:IET Cyber-Physical Systems
Subjects:
Online Access:https://doi.org/10.1049/cps2.12065
_version_ 1797263682984476672
author Hongjin Yao
Yisheng Li
Yunpeng Sun
Zhichao Lian
author_facet Hongjin Yao
Yisheng Li
Yunpeng Sun
Zhichao Lian
author_sort Hongjin Yao
collection DOAJ
description Abstract Recent work has shown that deep reinforcement learning (DRL) is vulnerable to adversarial attacks, so that exploiting vulnerabilities in DRL systems through adversarial attack techniques has become a necessary prerequisite for building robust DRL systems. Compared to traditional deep learning systems, DRL systems are characterised by long sequential decisions rather than one‐step decision, so attackers must perform multi‐step attacks on them. To successfully attack a DRL system, the number of attacks must be minimised to avoid detecting by the victim agent and to ensure the effectiveness of the attack. Some selective attack methods proposed in recent researches, that is, attacking an agent at partial time steps, are not applicable to real‐time attack scenarios, although they can avoid detecting by the victim agent. A real‐time selective attack method that is applicable to environments with discrete action spaces is proposed. Firstly, the optimal attack threshold T for performing selective attacks in the environment Env is determined. Then, the observation states corresponding to when the value of the action preference function of the victim agent in multiple eposides exceeds the threshold T are added to the training set according to this threshold. Finally, a universal perturbation is generated based on this training set, and it is used to perform real‐time selective attacks on the victim agent. Comparative experiments show that our attack method can perform real‐time attacks while maintaining the attack effect and stealthiness.
first_indexed 2024-04-25T00:16:54Z
format Article
id doaj.art-a5cf5495005c41dfbcf146138605fbce
institution Directory Open Access Journal
issn 2398-3396
language English
last_indexed 2024-04-25T00:16:54Z
publishDate 2024-03-01
publisher Wiley
record_format Article
series IET Cyber-Physical Systems
spelling doaj.art-a5cf5495005c41dfbcf146138605fbce2024-03-13T03:51:42ZengWileyIET Cyber-Physical Systems2398-33962024-03-0191414910.1049/cps2.12065Selective real‐time adversarial perturbations against deep reinforcement learning agentsHongjin Yao0Yisheng Li1Yunpeng Sun2Zhichao Lian3School of Computer Science and Technology Nanjing University of Science and Technology Nanjing ChinaSchool of Cyberspace Security Nanjing University of Science and Technology Nanjing ChinaSchool of Cyberspace Security Nanjing University of Science and Technology Nanjing ChinaSchool of Cyberspace Security Nanjing University of Science and Technology Nanjing ChinaAbstract Recent work has shown that deep reinforcement learning (DRL) is vulnerable to adversarial attacks, so that exploiting vulnerabilities in DRL systems through adversarial attack techniques has become a necessary prerequisite for building robust DRL systems. Compared to traditional deep learning systems, DRL systems are characterised by long sequential decisions rather than one‐step decision, so attackers must perform multi‐step attacks on them. To successfully attack a DRL system, the number of attacks must be minimised to avoid detecting by the victim agent and to ensure the effectiveness of the attack. Some selective attack methods proposed in recent researches, that is, attacking an agent at partial time steps, are not applicable to real‐time attack scenarios, although they can avoid detecting by the victim agent. A real‐time selective attack method that is applicable to environments with discrete action spaces is proposed. Firstly, the optimal attack threshold T for performing selective attacks in the environment Env is determined. Then, the observation states corresponding to when the value of the action preference function of the victim agent in multiple eposides exceeds the threshold T are added to the training set according to this threshold. Finally, a universal perturbation is generated based on this training set, and it is used to perform real‐time selective attacks on the victim agent. Comparative experiments show that our attack method can perform real‐time attacks while maintaining the attack effect and stealthiness.https://doi.org/10.1049/cps2.12065decision makinggradient methodsimage classification
spellingShingle Hongjin Yao
Yisheng Li
Yunpeng Sun
Zhichao Lian
Selective real‐time adversarial perturbations against deep reinforcement learning agents
IET Cyber-Physical Systems
decision making
gradient methods
image classification
title Selective real‐time adversarial perturbations against deep reinforcement learning agents
title_full Selective real‐time adversarial perturbations against deep reinforcement learning agents
title_fullStr Selective real‐time adversarial perturbations against deep reinforcement learning agents
title_full_unstemmed Selective real‐time adversarial perturbations against deep reinforcement learning agents
title_short Selective real‐time adversarial perturbations against deep reinforcement learning agents
title_sort selective real time adversarial perturbations against deep reinforcement learning agents
topic decision making
gradient methods
image classification
url https://doi.org/10.1049/cps2.12065
work_keys_str_mv AT hongjinyao selectiverealtimeadversarialperturbationsagainstdeepreinforcementlearningagents
AT yishengli selectiverealtimeadversarialperturbationsagainstdeepreinforcementlearningagents
AT yunpengsun selectiverealtimeadversarialperturbationsagainstdeepreinforcementlearningagents
AT zhichaolian selectiverealtimeadversarialperturbationsagainstdeepreinforcementlearningagents