From Semantics to Execution:Integrating Action Planning With Reinforcement Learning for Robotic Causal Problem-Solving
- Link:
- Autor/in:
- Erscheinungsjahr:
- 2019
- Medientyp:
- Text
- Schlagworte:
-
- causal puzzles
- hierarchical architecture
- neural networks
- planning
- reinforcement learning
- robotics
- Beschreibung:
-
-
Reinforcement learning is generally accepted to be an appropriate and successful method to learn robot control. Symbolic action planning is useful to resolve causal dependencies and to break a causally complex problem down into a sequence of simpler high-level actions. A problem with the integration of both approaches is that action planning is based on discrete high-level action- and state spaces, whereas reinforcement learning is usually driven by a continuous reward function. Recent advances in model-free reinforcement learning, specifically, universal value function approximators and hindsight experience replay, have focused on goal-independent methods based on sparse rewards that are only given at the end of a rollout, and only if the goal has been fully achieved. In this article, we build on these novel methods to facilitate the integration of action planning with model-free reinforcement learning. Specifically, the paper demonstrates how the reward-sparsity can serve as a bridge between the high-level and low-level state- and action spaces. As a result, we demonstrate that the integrated method is able to solve robotic tasks that involve non-trivial causal dependencies under noisy conditions, exploiting both data and knowledge.
-
- Lizenz:
-
- info:eu-repo/semantics/openAccess
- Quellsystem:
- Forschungsinformationssystem der UHH
Interne Metadaten
- Quelldatensatz
- oai:www.edit.fis.uni-hamburg.de:publications/1c8133f2-e1fd-42bb-8d79-ca9410daf2de