Gradient based algorithms with loss functions and kernels for improved on-policy control

Matthew Robards*, Peter Sunehag

*Corresponding author for this work

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    Abstract

    We introduce and empirically evaluate two novel online gradient-based reinforcement learning algorithms with function approximation - one model based, and the other model free. These algorithms come with the possibility of having non-squared loss functions which is novel in reinforcement learning, and seems to come with empirical advantages. We further extend a previous gradient based algorithm to the case of full control, by using generalized policy iteration. Theoretical properties of these algorithms are studied in a companion paper.

    Original languageEnglish
    Title of host publicationRecent Advances in Reinforcement Learning - 9th European Workshop, EWRL 2011, Revised Selected Papers
    Pages30-41
    Number of pages12
    DOIs
    Publication statusPublished - 2012
    Event9th European Workshop on Reinforcement Learning, EWRL 2011 - Athens, Greece
    Duration: 9 Sept 201111 Sept 2011

    Publication series

    NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
    Volume7188 LNAI
    ISSN (Print)0302-9743
    ISSN (Electronic)1611-3349

    Conference

    Conference9th European Workshop on Reinforcement Learning, EWRL 2011
    Country/TerritoryGreece
    CityAthens
    Period9/09/1111/09/11

    Fingerprint

    Dive into the research topics of 'Gradient based algorithms with loss functions and kernels for improved on-policy control'. Together they form a unique fingerprint.

    Cite this