RMIT University
Browse

Learning options for an MDP from demonstrations

conference contribution
posted on 2024-10-31, 18:41 authored by Marco Tamassia, Fabio ZambettaFabio Zambetta, William Raffe, Xiaodong LiXiaodong Li
The options framework provides a foundation to use hierarchical actions in reinforcement learning. An agent using options, along with primitive actions, at any point in time can decide to perform a macro-action made out of many primitive actions rather than a primitive action. Such macro-actions can be hand-crafted or learned. There has been previous work on learning them by exploring the environment. Here we take a different perspective and present an approach to learn options from a set of experts demonstrations. Empirical results are also presented in a similar setting to the one used in other works in this area.

History

Related Materials

  1. 1.
    DOI - Is published in 10.1007/978-3-319-14803-8_18
  2. 2.
    ISSN - Is published in 03029743

Start page

226

End page

242

Total pages

17

Outlet

Artificial Life and Computational Intelligence

Editors

Stephan K. Chalup, Alan D. Blair, and Marcus Randall

Name of conference

First Australasian Conference on Artificial Life and Computational Intelligence

Publisher

Springer International Publishing

Place published

Switzerland

Start date

2015-02-05

End date

2015-02-07

Language

English

Copyright

© 2015 Springer International Publishing Switzerland

Former Identifier

2006052195

Esploro creation date

2020-06-22

Fedora creation date

2015-04-20