Visible to the public Adversarial Examples Construction Towards White-Box Q Table Variation in DQN Pathfinding Training

TitleAdversarial Examples Construction Towards White-Box Q Table Variation in DQN Pathfinding Training
Publication TypeConference Paper
Year of Publication2018
AuthorsBai, X., Niu, W., Liu, J., Gao, X., Xiang, Y., Liu, J.
Conference Name2018 IEEE Third International Conference on Data Science in Cyberspace (DSC)
Keywordsadversarial examples, adversarial examples construction, adversarial samples, artificial intelligence, composability, countermeasure application scenario, deep reinforcement learning, DQN, DQN algorithm attack, DQN pathfinding training, learning (artificial intelligence), machine learning, Metrics, optimal path finding, Pathfinding, Prediction algorithms, pubcrawl, representative Deep Q Network algorithm, research hotspot, resilience, robotic automatic pathfinding application, robots, security, Task Analysis, Training, White Box Security, White-Box attack, White-box Q table variation
Abstract

As a new research hotspot in the field of artificial intelligence, deep reinforcement learning (DRL) has achieved certain success in various fields such as robot control, computer vision, natural language processing and so on. At the same time, the possibility of its application being attacked and whether it have a strong resistance to strike has also become a hot topic in recent years. Therefore, we select the representative Deep Q Network (DQN) algorithm in deep reinforcement learning, and use the robotic automatic pathfinding application as a countermeasure application scenario for the first time, and attack DQN algorithm against the vulnerability of the adversarial samples. In this paper, we first use DQN to find the optimal path, and analyze the rules of DQN pathfinding. Then, we propose a method that can effectively find vulnerable points towards White-Box Q table variation in DQN pathfinding training. Finally, we build a simulation environment as a basic experimental platform to test our method, through multiple experiments, we can successfully find the adversarial examples and the experimental results show that the supervised method we proposed is effective.

URLhttps://ieeexplore.ieee.org/document/8411947
DOI10.1109/DSC.2018.00126
Citation Keybai_adversarial_2018