Symbolic dynamic programming for discrete and continuous state MDPs

Scott Sanner*, Karina Valdivia Delgado, Leliane Nunes De Barros

*Corresponding author for this work

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    37 Citations (Scopus)

    Abstract

    Many real-world decision-theoretic planning problems can be naturally modeled with discrete and continuous state Markov decision processes (DC-MDPs). While previous work has addressed automated decision-theoretic planning for DCMDPs, optimal solutions have only been defined so far for limited settings, e.g., DC-MDPs having hyper-rectangular piecewise linear value functions. In this work, we extend symbolic dynamic programming (SDP) techniques to provide optimal solutions for a vastly expanded class of DCMDPs. To address the inherent combinatorial aspects of SDP, we introduce the XADD - a continuous variable extension of the algebraic decision diagram (ADD) - that maintains compact representations of the exact value function. Empirically, we demonstrate an implementation of SDP with XADDs on various DC-MDPs, showing the first optimal automated solutions to DCMDPs with linear and nonlinear piecewise partitioned value functions and showing the advantages of constraint-based pruning for XADDs.

    Original languageEnglish
    Title of host publicationProceedings of the 27th Conference on Uncertainty in Artificial Intelligence, UAI 2011
    PublisherAUAI Press
    Pages643-652
    Number of pages10
    Publication statusPublished - 2011

    Publication series

    NameProceedings of the 27th Conference on Uncertainty in Artificial Intelligence, UAI 2011

    Fingerprint

    Dive into the research topics of 'Symbolic dynamic programming for discrete and continuous state MDPs'. Together they form a unique fingerprint.

    Cite this