POMDP solution methods: a survey
2003

Darius Braziunas
Department of Computer Science
University of Toronto
Toronto, ON M5S 3H5


Abstract

This is an overview of partially observable Markov decision processes (POMDPs). We describe POMDP value and policy iteration as well as gradient ascent algorithms. The emphasis is on solution methods that work directly in the space of policies.

@TechReport{Braziunas-POMDPsurvey,
  author = 	 {Darius Braziunas},
  title = 	 {{POMDP} solution methods: a survey},
  institution =  {Department of Computer Science, University of Toronto},
  year = 	 {2003}
}


Download

To first page