Optimal use of experience in first person shooter environments

Matthew Aitchison*

*Corresponding author for this work

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    2 Citations (Scopus)

    Abstract

    Although reinforcement learning has made great strides recently, a continuing limitation is that it requires an extremely high number of interactions with the environment. In this paper, we explore the effectiveness of reusing experience from the experience replay buffer in the Deep Q-Learning algorithm. We test the effectiveness of applying learning update steps multiple times per environmental step in the VizDoom environment and show first, this requires a change in the learning rate, and second that it does not improve the performance of the agent. Furthermore, we show that updating less frequently is effective up to a ratio of 4:1, after which performance degrades significantly. These results quantitatively confirm the widespread practice of performing learning updates every 4th environmental step.

    Original languageEnglish
    Title of host publicationIEEE Conference on Games 2019, CoG 2019
    PublisherIEEE Computer Society
    ISBN (Electronic)9781728118840
    DOIs
    Publication statusPublished - Aug 2019
    Event2019 IEEE Conference on Games, CoG 2019 - London, United Kingdom
    Duration: 20 Aug 201923 Aug 2019

    Publication series

    NameIEEE Conference on Computatonal Intelligence and Games, CIG
    Volume2019-August
    ISSN (Print)2325-4270
    ISSN (Electronic)2325-4289

    Conference

    Conference2019 IEEE Conference on Games, CoG 2019
    Country/TerritoryUnited Kingdom
    CityLondon
    Period20/08/1923/08/19

    Fingerprint

    Dive into the research topics of 'Optimal use of experience in first person shooter environments'. Together they form a unique fingerprint.

    Cite this