Unknown

Dataset Information

0

From Semantics to Execution: Integrating Action Planning With Reinforcement Learning for Robotic Causal Problem-Solving.


ABSTRACT: Reinforcement learning is generally accepted to be an appropriate and successful method to learn robot control. Symbolic action planning is useful to resolve causal dependencies and to break a causally complex problem down into a sequence of simpler high-level actions. A problem with the integration of both approaches is that action planning is based on discrete high-level action- and state spaces, whereas reinforcement learning is usually driven by a continuous reward function. Recent advances in model-free reinforcement learning, specifically, universal value function approximators and hindsight experience replay, have focused on goal-independent methods based on sparse rewards that are only given at the end of a rollout, and only if the goal has been fully achieved. In this article, we build on these novel methods to facilitate the integration of action planning with model-free reinforcement learning. Specifically, the paper demonstrates how the reward-sparsity can serve as a bridge between the high-level and low-level state- and action spaces. As a result, we demonstrate that the integrated method is able to solve robotic tasks that involve non-trivial causal dependencies under noisy conditions, exploiting both data and knowledge.

SUBMITTER: Eppe M 

PROVIDER: S-EPMC7805615 | biostudies-literature | 2019

REPOSITORIES: biostudies-literature

altmetric image

Publications

From Semantics to Execution: Integrating Action Planning With Reinforcement Learning for Robotic Causal Problem-Solving.

Eppe Manfred M   Nguyen Phuong D H PDH   Wermter Stefan S  

Frontiers in robotics and AI 20191126


Reinforcement learning is generally accepted to be an appropriate and successful method to learn robot control. Symbolic action planning is useful to resolve causal dependencies and to break a causally complex problem down into a sequence of simpler high-level actions. A problem with the integration of both approaches is that action planning is based on <i>discrete high-level action- and state spaces</i>, whereas reinforcement learning is usually driven by a <i>continuous reward</i> function. Re  ...[more]

Similar Datasets

| S-EPMC6567809 | biostudies-literature
| S-EPMC6029810 | biostudies-literature
| S-EPMC3103422 | biostudies-literature
| S-EPMC5641385 | biostudies-other
| S-EPMC7695837 | biostudies-literature
| S-EPMC4523725 | biostudies-literature
| S-EPMC8606407 | biostudies-literature
| S-EPMC7979646 | biostudies-literature
| PRJEB40949 | ENA
| S-EPMC2749757 | biostudies-literature