16
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Open-Ended Learning: A Conceptual Framework Based on Representational Redescription

      brief-report

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Reinforcement learning (RL) aims at building a policy that maximizes a task-related reward within a given domain. When the domain is known, i.e., when its states, actions and reward are defined, Markov Decision Processes (MDPs) provide a convenient theoretical framework to formalize RL. But in an open-ended learning process, an agent or robot must solve an unbounded sequence of tasks that are not known in advance and the corresponding MDPs cannot be built at design time. This defines the main challenges of open-ended learning: how can the agent learn how to behave appropriately when the adequate states, actions and rewards representations are not given? In this paper, we propose a conceptual framework to address this question. We assume an agent endowed with low-level perception and action capabilities. This agent receives an external reward when it faces a task. It must discover the state and action representations that will let it cast the tasks as MDPs in order to solve them by RL. The relevance of the action or state representation is critical for the agent to learn efficiently. Considering that the agent starts with a low level, task-agnostic state and action spaces based on its low-level perception and action capabilities, we describe open-ended learning as the challenge of building the adequate representation of states and actions, i.e., of redescribing available representations. We suggest an iterative approach to this problem based on several successive Representational Redescription processes, and highlight the corresponding challenges in which intrinsic motivations play a key role.

          Related collections

          Most cited references18

          • Record: found
          • Abstract: not found
          • Conference Proceedings: not found

          Curriculum learning

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning

              Bookmark
              • Record: found
              • Abstract: not found
              • Article: not found

              Reinforcement learning in robotics: A survey

                Bookmark

                Author and article information

                Contributors
                Journal
                Front Neurorobot
                Front Neurorobot
                Front. Neurorobot.
                Frontiers in Neurorobotics
                Frontiers Media S.A.
                1662-5218
                25 September 2018
                2018
                : 12
                : 59
                Affiliations
                [1] 1Sorbonne Université, CNRS, ISIR , Paris, France
                [2] 2U2IS, INRIA Flowers, ENSTA ParisTech , Palaiseau, France
                [3] 3Institute of Perception, Action and Behaviour, University of Edinburgh , Edinburgh, United Kingdom
                [4] 4GII, Universidade da Coruña , A Coruña, Spain
                [5] 5Department of Computer Science, Vrije Universiteit Amsterdam , Amsterdam, Netherlands
                Author notes

                Edited by: Andrew Barto, University of Massachusetts Amherst, United States

                Reviewed by: Robert J. Lowe, University of Gothenburg, Sweden; Georg Martius, Max-Planck-Institut für Intelligente Systeme, Germany; Eiji Uchibe, Advanced Telecommunications Research Institute International (ATR), Japan

                *Correspondence: Stephane Doncieux stephane.doncieux@ 123456sorbonne-universite.fr
                Article
                10.3389/fnbot.2018.00059
                6167466
                847a7920-45e1-4fd9-9dfd-1a45da91e163
                Copyright © 2018 Doncieux, Filliat, Díaz-Rodríguez, Hospedales, Duro, Coninx, Roijers, Girard, Perrin and Sigaud.

                This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

                History
                : 28 April 2018
                : 28 August 2018
                Page count
                Figures: 2, Tables: 0, Equations: 0, References: 23, Pages: 6, Words: 4055
                Funding
                Funded by: H2020 Future and Emerging Technologies 10.13039/100010664
                Award ID: 640891
                Categories
                Neuroscience
                Perspective

                Robotics
                developmental robotics,reinforcement learning,state representation learning,representational redescription,actions and goals,skills

                Comments

                Comment on this article