ICAPS 04 Logo 14Th International Conference on Automated Planning & Scheduling

Whistler, British Columbia, Canada, June 3-7 2004
 
Welcome to ICAPS 2004
Registration Quicklink
Accommodation Quicklink
Author Instructions Quicklink
Program Quicklink
Calls
Events
Tutorials
Workshops
Doctoral Consortium
Co-Located
Pre-Conference
Program
People
 

Tutorial 5. Planning using Partially Observable Markov Decision Processes

Topic

Real-world planning problems are often characterized by partial observability, and there is increasing interest among planning researchers in developing planning algorithms that can select a proper course of action in spite of imperfect state information. This tutorial provides an introduction to research on planning using partially observable Markov decision processes (POMDPs), a very general and increasingly popular framework for planning under partial observability. The focus of the tutorial is on planning algorithms for POMDPs (in contrast to reinforcement learning approaches), and one of the goals is to show the continuity between traditional planning research and research on planning using POMDPs. In particular, coping with the challenge of partial observability requires development of techniques of knowledge representation, abstraction, problem decomposition, and heuristic search that generalize similar techniques used by traditional planners.

Intended Audience

No previous knowledge of Markov decision processes is assumed, and the tutorial will be accessible to anyone with basic familiarity with AI planning. Those who attend will acquire an understanding of the central concepts and algorithms for POMDPs, as well as a broad overview of recent research developments and applications.

Presenter Information

Eric Hansen Eric Hansen is an Associate Professor of Computer Science and Engineering at Mississippi State University. He received a A.B. in Philosophy from Harvard University, and a Ph.D. in Computer Science from the University of Massachusetts at Amherst in 1998. His PhD dissertation introduced a policy iteration algorithm for POMDPs that represents policies as finite-state controllers, and he has published several methods for scaling up algorithms for solving POMDPs. Other research interests include heuristic search and use of model-checking techniques in search and planning. He received an NSF Career Award in 2000.
Daniel Bernstein Daniel Bernstein is currently working towards a Ph.D. in Computer Science at the University of Massachusetts at Amherst. He received a B.S degree in Computer Science from Cornell University and an M.S. degree from the University of Massachusetts. His research interests include planning under uncertainty, reinforcement learning, and resource-bounded reasoning. His dissertation work focuses on planning for distributed agents in stochastic environments. Daniel is a recipient of an NSF Graduate Research Fellowship and a NASA GSRP Fellowship.
Zhengzhu Feng Zhengzhu Feng is a PhD student in the Computer Science Department at the University of Massachusetts at Amherst. He received a B.S. degree in Applied Mathematics from the South China University of Technology and an M.S. degree in Computer Science from Mississippi State University. His research interests include MDPs and POMDPs, and especially use of various forms of problem structure in scaling up planning algorithms for these models.
Rong Zhou Rong Zhou is a PhD student in Computer Science at Mississippi State University. He received a B.S degree in Electrical Engineering from Xi'an Jiaotong University, an M.S degree in Electrical Engineering from Tongji University, and an M.S. degree in Computer Science from Mississippi State University. His research interests include memory-efficient graph-search algorithms, as well as planning algorithms for POMDPs

Tutorial URL: http://www.cse.msstate.edu/~pomdp/ICAPS04Tutorial5.htm

Tutorial 1 Tutorial 2 Tutorial 3 Tutorial 4 Tutorial 5

Last modified: Tue Jun 15 08:59:13 Eastern Daylight Time 2004