Planning for Markov Decision Processes with Sparse Stochasticity - Robotics Institute Carnegie Mellon University

Planning for Markov Decision Processes with Sparse Stochasticity

Maxim Likhachev, Geoff Gordon, and Sebastian Thrun
Conference Paper, Proceedings of (NeurIPS) Neural Information Processing Systems, pp. 785 - 792, December, 2004

Abstract

Planning algorithms designed for deterministic worlds, such as A* search, usually run much faster than algorithms designed for worlds with uncertain action outcomes, such as value iteration. Real-world planning problems often exhibit uncertainty, which forces us to use the slower algorithms to solve them. Many real-world planning problems exhibit sparseuncertainty: there are long sequences of deterministic actions which accomplish tasks like moving sensor platforms into place, interspersed with a small number of sensing actions which have uncertain outcomes. In this paper we describe a new planning algorithm, called MCP (short for MDP Compression Planning), which combines A* search with value iteration for solving Stochastic Shortest Path problem in MDPs with sparse stochasticity. We present experiments which show that MCP can run substantially faster than competing planners in domains with sparse uncertainty; these experiments are based on a simulation of a ground robot cooperating with a helicopter to fill in a partial map and move to a goal location.

BibTeX

@conference{Likhachev-2004-109737,
author = {Maxim Likhachev and Geoff Gordon and Sebastian Thrun},
title = {Planning for Markov Decision Processes with Sparse Stochasticity},
booktitle = {Proceedings of (NeurIPS) Neural Information Processing Systems},
year = {2004},
month = {December},
pages = {785 - 792},
}