Temporal Logic Control of POMDPs via Label-based Stochastic Simulation Relations
From Murray WikiJump to navigationJump to search
|Temporal Logic Control of POMDPs via Label-based Stochastic Simulation Relations
|S. Haesaert, P. Nilsson, C. I. Vasile, R. Thakker, A. Agha-mohammadi and A. D. Ames and R. M. Murray
|To appear in IFAC Conference on Analysis and Design of Hybrid Systems (ADHS), 2018
|The synthesis of controllers guaranteeing linear temporal logic specifications on partially observable Markov decision processes (POMDP) via their belief models causes computational issues due to the continuous spaces. In this work, we construct a finite-state abstraction on which a control policy is synthesized and refined back to the original belief model. We introduce a new notion of label- based approximate stochastic simulation to quantify the deviation between belief models. We develop a robust synthesis methodology that yields a lower bound on the satisfaction probability, by compensating for deviations a priori, and that utilizes a less conservative control refinement.