Qitong GaoMiroslav PajićMichael M. Zavlanos
In this paper, we propose a Deep Imitative Q-learning (DIQL) method to synthesize control policies for mobile robots that need to satisfy Linear Temporal Logic (LTL) specifications using noisy semantic observations of their surroundings. The robot sensing error is modeled using probabilistic labels defined over the states of a Labeled Transition System (LTS) and the robot mobility is modeled using a Labeled Markov Decision Process (LMDP) with unknown transition probabilities. We use existing product-based model checkers (PMCs) as experts to guide the Q-learning algorithm to convergence. To the best of our knowledge, this is the first approach that models noise in semantic observations using probabilistic labeling functions and employs existing model checkers to provide suboptimal instructions to the Q-learning agent.
Mingyu CaiMohammadhosein HasanbeigShaoping XiaoAlessandro AbateZhen Kan
Nicola, GiorgioPedrocchi, NicolaGhidoni, Stefano
Nicola, GiorgioPedrocchi, NicolaGhidoni, Stefano
Wenbing TangFenghua WuShang‐Wei LinZuohua DingJing LiuYang LiuJifeng He