RMIT University
Browse

Deriving Subgoals Autonomously to Accelerate Learning in Sparse Reward Domains.

conference contribution
posted on 2024-10-31, 21:11 authored by Michael Dann, Fabio ZambettaFabio Zambetta, John ThangarajahJohn Thangarajah
Sparse reward games, such as the infamous Montezuma’s Revenge, pose a significant challenge for Reinforcement Learning (RL) agents. Hierarchical RL, which promotes efficient exploration via subgoals, has shown promise in these games. However, existing agents rely either on human domain knowledge or slow autonomous methods to derive suitable subgoals. In this work, we describe a new, autonomous approach for deriving subgoals from raw pixels that is more efficient than competing methods. We propose a novel intrinsic reward scheme for exploiting the derived subgoals, applying it to three Atari games with sparse rewards. Our agent’s performance is comparable to that of state-of-the-art methods, demonstrating the usefulness of the subgoals found.

History

Related Materials

  1. 1.
    DOI - Is published in 10.1609/aaai.v33i01.3301881
  2. 2.
    ISBN - Is published in 9781577358091 (urn:isbn:9781577358091)

Start page

881

End page

889

Total pages

9

Outlet

Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence (AAAI 2019)

Name of conference

AAAI 2019

Publisher

AAAI Press

Place published

California, United States

Start date

2019-01-27

End date

2019-02-01

Language

English

Copyright

Copyright © 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.

Former Identifier

2006094756

Esploro creation date

2020-06-22

Fedora creation date

2019-10-23

Usage metrics

    Scholarly Works

    Keywords

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC