A definition of happiness for reinforcement learning agents

Mayank Daswani, Jan Leike*

*Corresponding author for this work

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    5 Citations (Scopus)

    Abstract

    What is happiness for reinforcement learning agents? We seek a formal definition satisfying a list of desiderata. Our proposed definition of happiness is the temporal difference error, i.e. the difference between the value of the obtained reward and observation and the agent’s expectation of this value. This definition satisfies most of our desiderata and is compatible with empirical research on humans. We state several implications and discuss examples.

    Original languageEnglish
    Title of host publicationArtificial General Intelligence - 8th International Conference, AGI 2015, Proceedings
    EditorsAlexey Potapov, Jordi Bieger, Ben Goertzel
    PublisherSpringer Verlag
    Pages231-240
    Number of pages10
    ISBN (Print)9783319213644
    DOIs
    Publication statusPublished - 2015
    Event8th International Conference on Artificial General Intelligence, AGI 2015 - Berlin, Germany
    Duration: 22 Jul 201525 Jul 2015

    Publication series

    NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
    Volume9205
    ISSN (Print)0302-9743
    ISSN (Electronic)1611-3349

    Conference

    Conference8th International Conference on Artificial General Intelligence, AGI 2015
    Country/TerritoryGermany
    CityBerlin
    Period22/07/1525/07/15

    Fingerprint

    Dive into the research topics of 'A definition of happiness for reinforcement learning agents'. Together they form a unique fingerprint.

    Cite this