TY - JOUR
T1 - Adversarial Jamming Attacks and Defense Strategies via Adaptive Deep Reinforcement Learning
AU - Wang, Feng
AU - Zhong, Chen
AU - Gursoy, M. Cenk
AU - Velipasalar, Senem
N1 - Publisher Copyright:
Copyright © 2020, The Authors. All rights reserved.
Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
PY - 2020/7/12
Y1 - 2020/7/12
N2 - —As the applications of deep reinforcement learning (DRL) in wireless communications grow, sensitivity of DRL based wireless communication strategies against adversarial attacks has started to draw increasing attention. In order to address such sensitivity and alleviate the resulting security concerns, we in this paper consider a victim user that performs DRL-based dynamic channel access, and an attacker that executes DRL-based jamming attacks to disrupt the victim. Hence, both the victim and attacker are DRL agents and can interact with each other, retrain their models, and adapt to opponents’ policies. In this setting, we initially develop an adversarial jamming attack policy that aims at minimizing the accuracy of victim’s decision making on dynamic channel access. Subsequently, we devise defense strategies against such an attacker, and propose three defense strategies, namely diversified defense with proportional-integral-derivative (PID) control, diversified defense with an imitation attacker, and defense via orthogonal policies. We design these strategies to maximize the attacked victims accuracy and evaluate their performances.MSC Codes 68T07 (Primary) 94A15 (Secondary)
AB - —As the applications of deep reinforcement learning (DRL) in wireless communications grow, sensitivity of DRL based wireless communication strategies against adversarial attacks has started to draw increasing attention. In order to address such sensitivity and alleviate the resulting security concerns, we in this paper consider a victim user that performs DRL-based dynamic channel access, and an attacker that executes DRL-based jamming attacks to disrupt the victim. Hence, both the victim and attacker are DRL agents and can interact with each other, retrain their models, and adapt to opponents’ policies. In this setting, we initially develop an adversarial jamming attack policy that aims at minimizing the accuracy of victim’s decision making on dynamic channel access. Subsequently, we devise defense strategies against such an attacker, and propose three defense strategies, namely diversified defense with proportional-integral-derivative (PID) control, diversified defense with an imitation attacker, and defense via orthogonal policies. We design these strategies to maximize the attacked victims accuracy and evaluate their performances.MSC Codes 68T07 (Primary) 94A15 (Secondary)
KW - Adversarial policies
KW - Deep reinforcement learning
KW - Defense strategies
KW - Dynamic channel access
KW - Jamming attacks
UR - http://www.scopus.com/inward/record.url?scp=85095498619&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85095498619&partnerID=8YFLogxK
M3 - Article
AN - SCOPUS:85095498619
JO - Nuclear Physics A
JF - Nuclear Physics A
SN - 0375-9474
ER -