Inviting an author to review:
Find an author and click ‘Invite to review selected article’ near their name.
Search for authorsSearch for similar articles
1
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Conference Proceedings: not found

      Tactics of Adversarial Attack on Deep Reinforcement Learning Agents

      proceedings-article
      1 , 1 , 1 , 1 , 2 , 1
      Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI-2017)
      Artificial Intelligence
      September 19, 2017 - September 26, 2017

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          We introduce two tactics, namely the strategically-timed attack and the enchanting attack, to attack reinforcement learning agents trained by deep reinforcement learning algorithms using adversarial examples. In the strategically-timed attack, the adversary aims at minimizing the agent's reward by only attacking the agent at a small subset of time steps in an episode. Limiting the attack activity to this subset helps prevent detection of the attack by the agent. We propose a novel method to determine when an adversarial example should be crafted and applied. In the enchanting attack, the adversary aims at luring the agent to a designated target state. This is achieved by combining a generative model and a planning algorithm: while the generative model predicts the future states, the planning algorithm generates a preferred sequence of actions for luring the agent. A sequence of adversarial examples is then crafted to lure the agent to take the preferred sequence of actions. We apply the proposed tactics to the agents trained by the state-of-the-art deep reinforcement learning algorithm including DQN and A3C. In 5 Atari games, our strategically-timed attack reduces as much reward as the uniform attack (i.e., attacking at every time step) does by attacking the agent 4 times less often. Our enchanting attack lures the agent toward designated target states with a more than 70% success rate. Example videos are available at http://yclin.me/adversarial_attack_RL/.

          Related collections

          Author and article information

          Conference
          August 2017
          August 2017
          : 3756-3762
          Affiliations
          [1 ]National Tsing Hua University
          [2 ]Nvidia
          Article
          10.24963/ijcai.2017/525
          6a38c25d-284f-4e72-99f9-b53431a201d9
          © 2017
          Twenty-Sixth International Joint Conference on Artificial Intelligence
          IJCAI-2017
          26
          Melbourne, Australia
          September 19, 2017 - September 26, 2017
          International Joint Conferences on Artificial Intelligence Organization (IJCAI)
          University of Technology Sydney (UTS)
          Australian Computer Society (ACS)
          Artificial Intelligence
          History

          Comments

          Comment on this article