In Markov Decision Processes (MDPs), rewards are assigned according to a function of the last state and action. This is often limiting, when the considered domain is not naturally Markovian, but becomes so after careful engineering of extended state space. The extended states record information from the past that is sufficient to assign rewards by looking just at the last state and action. Non-Markovian Reward Decision Processes (NRMDPs) extend MDPs by allowing for non-Markovian rewards, which depend on the history of states and actions. Non-Markovian rewards can be specified in temporal logics on finite traces such as LTLf/LDLf, with the great advantage of a higher abstraction and succinctness; they can then be automatically compiled into an MDP with an extended state space. We contribute to the techniques to handle temporal rewards and to the solutions to engineer them. We first present an approach to compiling temporal rewards which merges the formula automata into a single transducer, sometimes saving up to an exponential number of states. We then define monitoring rewards, which add a further level of abstraction to temporal rewards by adopting the four-valued conditions of runtime monitoring; we argue that our compilation technique allows for an efficient handling of monitoring rewards. Finally, we discuss application to reinforcement learning.
Dettaglio pubblicazione
2020, Proceedings of the 17th International Conference on Principles of Knowledge Representation and Reasoning, Pages 860-870
Temporal Logic Monitoring Rewards via Transducers (04b Atto di convegno in volume)
De Giacomo Giuseppe, Favorito Marco, Iocchi Luca, Patrizi Fabio, Ronca Alessandro
ISBN: 978-0-9992411-7-2
Gruppo di ricerca: Artificial Intelligence and Knowledge Representation
keywords