Reinforcement learning with non-markovian rewards

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    67 Scopus citations

    Abstract

    The standard RL world model is that of a Markov Decision Process (MDP). A basic premise of MDPs is that the rewards depend on the last state and action only. Yet, many real-world rewards are non-Markovian. For example, a reward for bringing coffee only if requested earlier and not yet served, is non-Markovian if the state only records current requests and deliveries. Past work considered the problem of modeling and solving MDPs with non-Markovian rewards (NMR), but we know of no principled approaches for RL with NMR. Here, we address the problem of policy learning from experience with such rewards. We describe and evaluate empirically four combinations of the classical RL algorithm Q-learning and R-max with automata learning algorithms to obtain new RL algorithms for domains with NMR. We also prove that some of these variants converge to an optimal policy in the limit.

    Original languageEnglish
    Title of host publicationAAAI 2020 - 34th AAAI Conference on Artificial Intelligence
    PublisherAAAI press
    Pages3980-3987
    Number of pages8
    ISBN (Electronic)9781577358350
    StatePublished - 1 Jan 2020
    Event34th AAAI Conference on Artificial Intelligence, AAAI 2020 - New York, United States
    Duration: 7 Feb 202012 Feb 2020

    Publication series

    NameAAAI 2020 - 34th AAAI Conference on Artificial Intelligence

    Conference

    Conference34th AAAI Conference on Artificial Intelligence, AAAI 2020
    Country/TerritoryUnited States
    CityNew York
    Period7/02/2012/02/20

    ASJC Scopus subject areas

    • Artificial Intelligence

    Fingerprint

    Dive into the research topics of 'Reinforcement learning with non-markovian rewards'. Together they form a unique fingerprint.

    Cite this