Goal-HSVI: Heuristic Search Value Iteration for Goal POMDPs
Goal-HSVI: Heuristic Search Value Iteration for Goal POMDPs
Karel Horák, Branislav Bošanský, Krishnendu Chatterjee
Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence
Main track. Pages 4764-4770.
https://doi.org/10.24963/ijcai.2018/662
Partially observable Markov decision processes (POMDPs) are the standard models for planning under uncertainty with both finite and infinite horizon. Besides the well-known discounted-sum objective, indefinite-horizon objective (aka Goal-POMDPs) is another classical objective for POMDPs. In this case, given a set of target states and a positive cost for each transition, the optimization objective is to minimize the expected total cost until a target state is reached.
In the literature, RTDP-Bel or heuristic search value iteration (HSVI) have been used for solving Goal-POMDPs. Neither of these algorithms has theoretical convergence guarantees, and HSVI may even fail to terminate its trials. We give the following contributions: (1) We discuss the challenges introduced in Goal-POMDPs and illustrate how they prevent the original HSVI from converging. (2) We present a novel algorithm inspired by HSVI, termed Goal-HSVI, and show that our algorithm has convergence guarantees. (3) We show that Goal-HSVI outperforms RTDP-Bel on a set of well-known examples.
Keywords:
Planning and Scheduling: POMDPs
Planning and Scheduling: Theoretical Foundations of Planning
Planning and Scheduling: Planning under Uncertainty