When modeling real-world decision-theoretic planning problems in the Markov decision process (MDP) framework, it is often impossible to obtain a completely accurate estimate of transition probabilities. For example, natural uncertainty arises in the transition specification due to elicitation of MDP transition models from an expert or data, or non-stationary transition distributions arising from insufficient state knowledge. In the interest of obtaining the most robust policy under transition uncertainty, the Markov Decision Process with Imprecise Transition Probabilities (MDP-1Ps) has been introduced to model such scenarios. Unfortunately, while solutions to the MDP-1P are well-known, they require nonlinear optimization and are extremely time-consuming in practice. To address this deficiency, we propose efficient dynamic programming methods to exploit the structure of factored MDP-1Ps. Noting that the key computational bottleneck in the solution of MDP-1Ps is the need to repeatedly solve nonlinear constrained optimization problems, we show how to target approximation techniques to drastically reduce the computational overhead of the nonlinear solver while producing bounded, approximately optimal solutions. Our results show up to two orders of magnitude speedup in comparison to traditional "flat" dynamic programming approaches and up to an order of magnitude speedup over the extension of factored MDP approximate value iteration techniques to MDP-1Ps. Copyright © 2009, Association for the Advancement of Artificial Intelligence. All rights reserved.
|Original language||American English|
|Number of pages||8|
|State||Published - 1 Dec 2009|
|Event||ICAPS 2009 - Proceedings of the 19th International Conference on Automated Planning and Scheduling - |
Duration: 1 Dec 2009 → …
|Conference||ICAPS 2009 - Proceedings of the 19th International Conference on Automated Planning and Scheduling|
|Period||1/12/09 → …|