3
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Deep Reinforcement Learning Controller for 3D Path Following and Collision Avoidance by Autonomous Underwater Vehicles

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Control theory provides engineers with a multitude of tools to design controllers that manipulate the closed-loop behavior and stability of dynamical systems. These methods rely heavily on insights into the mathematical model governing the physical system. However, in complex systems, such as autonomous underwater vehicles performing the dual objective of path following and collision avoidance, decision making becomes nontrivial. We propose a solution using state-of-the-art Deep Reinforcement Learning (DRL) techniques to develop autonomous agents capable of achieving this hybrid objective without having a priori knowledge about the goal or the environment. Our results demonstrate the viability of DRL in path following and avoiding collisions towards achieving human-level decision making in autonomous vehicle systems within extreme obstacle configurations.

          Related collections

          Most cited references40

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          Curriculum learning

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            The dynamic window approach to collision avoidance

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Continuous control with deep reinforcement learning

              We adapt the ideas underlying the success of Deep Q-Learning to the continuous action domain. We present an actor-critic, model-free algorithm based on the deterministic policy gradient that can operate over continuous action spaces. Using the same learning algorithm, network architecture and hyper-parameters, our algorithm robustly solves more than 20 simulated physics tasks, including classic problems such as cartpole swing-up, dexterous manipulation, legged locomotion and car driving. Our algorithm is able to find policies whose performance is competitive with those found by a planning algorithm with full access to the dynamics of the domain and its derivatives. We further demonstrate that for many of the tasks the algorithm can learn policies end-to-end: directly from raw pixel inputs. 10 pages + supplementary
                Bookmark

                Author and article information

                Contributors
                Journal
                Front Robot AI
                Front Robot AI
                Front. Robot. AI
                Frontiers in Robotics and AI
                Frontiers Media S.A.
                2296-9144
                25 January 2021
                2020
                : 7
                : 566037
                Affiliations
                [ 1 ]Department of Engineering Cybernetics, Norwegian University of Science and Technology, Trondheim, Norway
                [ 2 ]Mathematics and Cybernetics, SINTEF Digital, Trondheim, Norway
                [ 3 ]School of Mechanical and Aerospace Engineering, Oklahoma State University, Stillwater, OK, United States
                Author notes

                Edited by: Holger Voos, University of Luxembourg, Luxembourg

                Reviewed by: Marcello Cirillo, Scania, Sweden

                Martin Ludvigsen, Norwegian University of Science and Technology, Norway

                *Correspondence: Adil Rasheed, adil.rasheed@ 123456ntnu.no

                This article was submitted to Robotic Control Systems, a section of the journal Frontiers in Robotics and AI

                Article
                566037
                10.3389/frobt.2020.566037
                7874127
                c4efb3b2-5c3e-4987-a5dc-9ffc8b3ae1aa
                Copyright © 2021 Havenstrøm, Rasheed and San.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

                History
                : 26 May 2020
                : 08 December 2020
                Categories
                Robotics and AI
                Original Research

                continuous control,collision avoidance,path following,deep reinforcement learning,autonomous under water vehicle,curriculum learning

                Comments

                Comment on this article