Planning Under Uncertainty
Classical Planning is Unrealistic
Nondeterminism is the Rule of Nature
Planning with Non-Deterministic Domains
Planning with Markov Decision Process
Planning with Markov Decision Process
Planning with Fully Observable Domains
Domain as Stochastic System
Nondeterministic Actions
move(r1,l2,l3)
move(r1,l1,l4)
Plan as Policies
Example Policies
All three policies try to move the robot toward state s4.
Policy Execution
Probability of a History
Goals as Utility Functions
l1
l2
l3
l5
l4
Goals as Utility Functions
Utility Function
Utility of a History
Expected Utility of a Policy
Example expected utility
Planning as an Optimization Problem
Expected Utility without Reward
When there are no rewards associated with the states
Bellman Equation
Planning Algorithm
Solve system of equations
Improve the policy by choosing
Optimal Policy:
Possible iterations for state s1 at the first step