POMDP solution methods: a survey

Darius Braziunas
Department of Computer Science
University of Toronto
Toronto, ON M5S 3H5


This is an overview of partially observable Markov decision processes (POMDPs). We describe POMDP value and policy iteration as well as gradient ascent algorithms. The emphasis is on solution methods that work directly in the space of policies.

  author = 	 {Darius Braziunas},
  title = 	 {{POMDP} solution methods: a survey},
  institution =  {Department of Computer Science, University of Toronto},
  year = 	 {2003}


To first page