Probabilistic Robotic Reading Note - Approximate POMDP Techniques

Subscribe Send me a message home page

As we saw in the post Probabilistic Robotic Reading Note - Partially Observable Markov Decision Processes, the complexity of the calculation of value function is exponential. For any real-world problem, the state space and the belif space are usually large enough to make the naive POMDP implementation impractical.

Chapter 16 of the book describes three approximate POMDP techniques

In this post, we will just copy the algorithm box from the book and add some notes.

Some general notes:

The effect of the uncertain is best illustrated by the following figure copied from the book. In a non probabilistic world (images on the left), where the state is known and there is no uncertainty, open space is obviously preferred by the planner because the robot is less likely to hit anything and crossing an open space usually produces a shorter path. However, if the state cannot be known with certainty (images on the right), then the robot will have a hard time to localize itself in an open space. It follows that the landmarks, including obstructions, in the environment are more valuable because they provide more information.


QMDP Algorithm


AMDP Algorithm


MC-POMDP Algorithm


Note that a belief is a collection of sampled states (e.g. particles) and the ultimate goal of the algorithm is to assign scores to thoses beliefs so we have

----- END -----

Send me a message Subscribe to blog updates

Want some fun stuff?