Abstract
We introduce and study Regular Decision Processes (RDPs), a new, compact model for domains with non-Markovian dynamics and rewards, in which the dependence on the past is regular, in the language theoretic sense. RDPs are an intermediate model between MDPs and POMDPs. They generalize k-order MDPs and can be viewed as a POMDP in which the hidden state is a regular function of the entire history. In factored RDPs, transition and reward functions are specified using formulas in linear temporal logics over finite traces, or using regular expressions. This allows specifying complex dependence on the past using intuitive and compact formulas, and building models of partially observable domains without specifying an underlying state space.
Original language | English |
---|---|
Article number | 104113 |
Journal | Artificial Intelligence |
Volume | 331 |
DOIs | |
State | Published - 1 Jun 2024 |
Keywords
- Markov-decision processes
- Non-Markovian decision processes
- POMDPs
- Regular languages
ASJC Scopus subject areas
- Language and Linguistics
- Linguistics and Language
- Artificial Intelligence