Temporal Logic Monitoring Rewards via Transducers
Autor: | Marco Favorito, Alessandro Ronca, Fabio Patrizi, Luca Iocchi, Giuseppe De Giacomo |
---|---|
Rok vydání: | 2020 |
Předmět: |
Computer science
0102 computer and information sciences 02 engineering and technology 01 natural sciences action languages Transducer 010201 computation theory & mathematics Symbolic reinforcement learning 0202 electrical engineering electronic engineering information engineering Electronic engineering 020201 artificial intelligence & image processing Temporal logic General Reasoning about actions and change |
Zdroj: | Proceedings of the Seventeenth International Conference on Principles of Knowledge Representation and Reasoning KR |
DOI: | 10.24963/kr.2020/89 |
Popis: | In Markov Decision Processes (MDPs), rewards are assigned according to a function of the last state and action. This is often limiting, when the considered domain is not naturally Markovian, but becomes so after careful engineering of extended state space. The extended states record information from the past that is sufficient to assign rewards by looking just at the last state and action. Non-Markovian Reward Decision Processes (NRMDPs) extend MDPs by allowing for non-Markovian rewards, which depend on the history of states and actions. Non-Markovian rewards can be specified in temporal logics on finite traces such as LTLf/LDLf, with the great advantage of a higher abstraction and succinctness; they can then be automatically compiled into an MDP with an extended state space. We contribute to the techniques to handle temporal rewards and to the solutions to engineer them. We first present an approach to compiling temporal rewards which merges the formula automata into a single transducer, sometimes saving up to an exponential number of states. We then define monitoring rewards, which add a further level of abstraction to temporal rewards by adopting the four-valued conditions of runtime monitoring; we argue that our compilation technique allows for an efficient handling of monitoring rewards. Finally, we discuss application to reinforcement learning. |
Databáze: | OpenAIRE |
Externí odkaz: |