Markov Decision Processes: Making Decision in the Presence of Uncertainty. Decision Processes: General Description

Similar documents
ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015

An Introduction to Markov Decision Processes. MDP Tutorial - 1

Reinforcement Learning

Preliminaries: Problem Definition Agent model, POMDP, Bayesian RL

Reinforcement Learning

Decision Making under Uncertainty

6.3 Conditional Probability and Independence

Lecture 1: Introduction to Reinforcement Learning

Math/Stats 425 Introduction to Probability. 1. Uncertainty and the axioms of probability

Objectives. Materials

Section 1.3 P 1 = 1 2. = P n = 1 P 3 = Continuing in this fashion, it should seem reasonable that, for any n = 1, 2, 3,..., =

The Clute Institute International Academic Conference Las Vegas, Nevada, USA 2013

Game playing. Chapter 6. Chapter 6 1

CHAPTER 2 Estimating Probabilities

cs171 HW 1 - Solutions

Linear Programming Notes V Problem Transformations

Measuring the Performance of an Agent

A survey of point-based POMDP solvers

AI: A Modern Approach, Chpts. 3-4 Russell and Norvig

6.254 : Game Theory with Engineering Applications Lecture 2: Strategic Form Games

Planning Treatment of Ischemic Heart Disease with Partially Observable Markov Decision Processes

Motivation. Motivation. Can a software agent learn to play Backgammon by itself? Machine Learning. Reinforcement Learning

(Refer Slide Time: 2:03)

Binomial lattice model for stock prices

Computational Geometry Lab: FEM BASIS FUNCTIONS FOR A TETRAHEDRON

Numerical Methods for Option Pricing

A Sarsa based Autonomous Stock Trading Agent

Math Quizzes Winter 2009

CHAPTER 7 APPLICATIONS TO MARKETING. Chapter 7 p. 1/54

6.231 Dynamic Programming and Stochastic Control Fall 2008

Financial Mathematics and Simulation MATH Spring 2011 Homework 2

Using Markov Decision Processes to Solve a Portfolio Allocation Problem

A Static Analyzer for Large Safety-Critical Software. Considered Programs and Semantics. Automatic Program Verification by Abstract Interpretation

Analysis of Micromouse Maze Solving Algorithms

Temporal Difference Learning in the Tetris Game

Lecture Note 1 Set and Probability Theory. MIT Spring 2006 Herman Bennett

Chapter 11 Monte Carlo Simulation

POMPDs Make Better Hackers: Accounting for Uncertainty in Penetration Testing. By: Chris Abbott

GCSE Business Studies. Ratios. For first teaching from September 2009 For first award in Summer 2011

1 The Brownian bridge construction

1 Interest rates, and risk-free investments

Dynamic Programming. Lecture Overview Introduction

NPV Versus IRR. W.L. Silber We know that if the cost of capital is 18 percent we reject the project because the NPV

Section 6.1 Discrete Random variables Probability Distribution

Investment, Time, and Present Value

Numeracy and mathematics Experiences and outcomes

Special Situations in the Simplex Algorithm

The Graphical Method: An Example

WRITING PROOFS. Christopher Heil Georgia Institute of Technology

Reinforcement Learning of Task Plans for Real Robot Systems

Geometric Series and Annuities

3.2 Roulette and Markov Chains

Tutorial Customer Lifetime Value

One Period Binomial Model

ST 371 (IV): Discrete Random Variables

Overview. Essential Questions. Precalculus, Quarter 4, Unit 4.5 Build Arithmetic and Geometric Sequences and Series

1 The EOQ and Extensions

Chapter 4 DECISION ANALYSIS

Skill-based Routing Policies in Call Centers

Decision Theory Rational prospecting

Math 120 Final Exam Practice Problems, Form: A

Time Value of Money 1

CS104: Data Structures and Object-Oriented Design (Fall 2013) October 24, 2013: Priority Queues Scribes: CS 104 Teaching Team

A PRELIMINARY REPORT ON A GENERAL THEORY OF INDUCTIVE INFERENCE

Data Structures and Algorithms

Linear Programming. Solving LP Models Using MS Excel, 18

10.2 Series and Convergence

Dynamic Programming 11.1 AN ELEMENTARY EXAMPLE

VENDOR MANAGED INVENTORY

Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 10

The fairy tale Hansel and Gretel tells the story of a brother and sister who

Linear Programming Notes VII Sensitivity Analysis

APP INVENTOR. Test Review

What is Linear Programming?

Language Modeling. Chapter Introduction

Sowing Games JEFF ERICKSON

Self Organizing Maps: Fundamentals

People have thought about, and defined, probability in different ways. important to note the consequences of the definition:

FREE FALL. Introduction. Reference Young and Freedman, University Physics, 12 th Edition: Chapter 2, section 2.5

Numbers 101: Cost and Value Over Time

Sequential lmove Games. Using Backward Induction (Rollback) to Find Equilibrium

HURDLE AND SELECTION MODELS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics July 2009

THERE is a significant amount of current research activity

Speech at IFAC2014 BACKGROUND

1 Review of Least Squares Solutions to Overdetermined Systems

STB- 2. Installation and Operation Manual

Optimal Design of Sequential Real-Time Communication Systems Aditya Mahajan, Member, IEEE, and Demosthenis Teneketzis, Fellow, IEEE

Overview. Data Mining. Predicting Stock Market Returns. Predicting Health Risk. Wharton Department of Statistics. Wharton

Introduction to MATLAB IAP 2008

Guessing Game: NP-Complete?

Computer Game Programming Intelligence I: Basic Decision-Making Mechanisms

Present Value, Discounted Cash Flow. Engineering Economy

1 Error in Euler s Method

Multiplication and Division with Rational Numbers

6.231 Dynamic Programming Midterm, Fall Instructions

Introduction to LAN/WAN. Network Layer

Math Games For Skills and Concepts

BrainMaster Macromedia Flash Player

The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy

Transcription:

Markov Decision Processes: Making Decision in the Presence of Uncertainty (some of) R&N 6.-6.6 R&N 7.-7.4 Decision Processes: General Description Suppose that you own a business. At any time, you know exactly the current state of the business (finances, stock, etc.). At any time, you have a choice of several possible actions (advertise, introduce new products, ). You cannot predict with certainty the consequence of these actions given the current state of the business, but you have a guess as to the likelihood of the possible outcomes. How can you define a policy that will guarantee that you always choose the action that maximizes expected future profits? Note: Russel & Norvig, Chapter 7.

Decision Processes: General Description Decide what action to take next, given: A probability to move to different states A way to evaluate the reward of being in different states Robot path planning Travel route planning Elevator scheduling Aircraft navigation Manufacturing processes Network switching & routing Example Assume that time is discretized into discrete time steps t =,2,. (for example, fiscal years) Suppose that the business can be in one of a finite number of states s (this is a major simplification, but let s assume.) Suppose that for every state s, we can anticipate a reward that the business receives for being in that state: R(s) (in this example, R(s) would the profit, possibly negative, generated by the business) Assume also that R(s) is bounded (R(s) < M for all s), meaning that the business cannot generate more than a certain profit threshold Question: What is the total value of the reward for a particular configuration of states {s,s 2, } over time? 2

Example Question: What is the total value of the reward for a particular configuration of states {s,s 2, } over time? It is simply the sum of the rewards (possibly negative) that we will receive in the future: U(s,s 2,.., s n,..) = R(s )+R(s 2 )+..+R(s n ) +... What is wrong with this formula??? Horizon Problem U(s 0,, s N ) = R(s 0 )+R(s )+ +R(s N ) The sum may be arbitrarily large depending on N Need to know N, the length of the sequence (finite horizon) 3

Horizon Problem The problem is that we did not put any limit on the future, so this sum can be infinite. For example: Consider the simple case of computing the total future reward if the business remains forever in the same state: U(s,s,.., s,..) = R(s)+R(s)+..+ R(s) +... is clearly infinite in general!! This definition is useless unless we consider a finite time horizon. But, in general, we don t have a good way to define such a time horizon. Discounting U(s 0,.)=R(s 0 )+γr(s )+..+γ N R(s N )+.. The length of the sequence is arbitrary (infinite horizon) Discount factor 0 < γ < 4

Discounting U(s 0,..) = R(s 0 ) + γr(s ) +.+ γ N R(s N ) +. Always converges if γ < and R(.) is bounded γ close to 0 instant gratification, don t pay attention to future reward γ close to extremely conservative, consider profits/losses no matter how far in the future The resulting model is the discounted reward Prefers expedient solutions (models impatience) Compensates for uncertainty in available time (models mortality) Economic example: Being promised $0,000 next year is worth only 90% as much as receiving $0,000 right now. Assuming payment n years in future is worth only (0.9) n of payment now Actions Assume that we also have a finite set of actions a An action a causes a transition from a state s to a state s In the business example, an action may be placing advertising, or selling stock, etc. 5

The Basic Decision Problem Given: Set of states S = {s} Set of actions A = {a} a: S S Reward function R(.) Discount factor γ Starting state s Find a sequence of actions such that the resulting sequence of states maximizes the total discounted reward: U(s 0,.)=R(s 0 )+γr(s )+..+γ N R(s N )+.. Maze Example: Utility 2 3 4 3 2 + - Define the reward of being in a state: R(s) = -0.04 if s is empty state R(4,3) = + (maximum reward when goal is reached) R(4,2) = - (avoid (4,2) as much as possible) Define the utility of a sequence of states: U(s 0,, s N ) = R(s 0 ) + R(s ) +.+R(s N ) 6

Maze Example: Utility 3 2 2 3 4 START If + no uncertainty: Find the sequence of actions - that maximizes the sum of the rewards of the traversed states Define the reward of being in a state: R(s) = -0.04 if s is empty state R(4,3) = + (maximum reward when goal is reached) R(4,2) = - (avoid (4,2) as much as possible) Define the utility of a sequence of states: U(s 0,, s N ) = R(s 0 ) + R(s ) +.+R(s N ) Maze Example: No Uncertainty 2 3 4 3 + 2 - START States: locations in maze grid Actions: Moves up/left left/right If no uncertainty: Find sequence of actions from current state to goal (+) that maximizes utility We know how to do this using earlier search techniques 7

What we are looking for: Policy Policy = Mapping from states to action π(s) = a Which action should I take in each state In the maze example, π(s) associates a motion to a particular location on the grid For any state s, we define the utility U(s) of s as the sum of discounted rewards of the sequence of states starting at state s generated by using the policy π U(s) = R(s) + γ R(s ) + γ 2 R(s 2 ) +.. Where we move from s to s by action π(s) We move from s to s 2 by action π(s ) etc. Optimal Decision Policy Policy = Mapping from states to action π(s) = a Which action should I take in each state Intuition: π encodes the best action that we can take from any state to maximize future rewards In the maze example, π(s) associates a motion to a particular location on the grid Optimal Policy = The policy π* that maximizes the expected utility U(s) of the sequence of states generated by π*, starting at s In the maze example, π*(s) tells us which motion to choose at every cell of the grid to bring us closer to the goal 8

Maze Example: No Uncertainty 2 3 4 3 2 + - START π*((,)) = UP π*((,3)) = RIGHT π*((4,)) = LEFT Maze Example: With Uncertainty Intended action: Executed action: Prob = 0.8 Prob = 0.0 Prob = 0. Prob = 0. The robot may not execute exactly the action that is commanded The outcome of an action is no longer deterministic Uncertainty: We know in which state we are (fully observable) But we are not sure that the commanded action will be executed exactly 9

Uncertainty No uncertainty: An action a deterministically causes a transition from a state s to another state s With uncertainty: An action a causes a transition from a state s to another state s with some probability T(s,a,s ) T(s,a,s ) is called the transition probability from state s to state s through action a In general, we need S 2 x A numbers to store all the transitions probabilities Maze Example: With Uncertainty 2 3 4 3 2 + - We can no longer find a unique sequence of actions, but Can we find a policy that tells us how to decide which action to take from each state except that now the policy maximizes the expected utility 0

3 2 Maze Example: Utility Revisited 2 3 4 + - Intended action a: 0. T(s,a,s ) 0.8 0. U(s) = Expected reward of future states starting at s How to compute U after one step? 3 2 Maze Example: Utility Revisited 2 3 4 s + - Intended action a: 0. T(s,a,s ) 0.8 0. Suppose s = (,) and we choose action Up. U(,) = R(,) +

3 2 Maze Example: Utility Revisited 2 3 4 s + - Intended action a: 0. T(s,a,s ) 0.8 0. Suppose s = (,) and we choose action Up. U(,) = R(,) + 0.8 x U(,2) + 3 2 Maze Example: Utility Revisited 2 3 4 s + - Intended action a: 0. T(s,a,s ) 0.8 0. Suppose s = (,) and we choose action Up. U(,) = R(,) + 0.8 x U(,2) + 0. x U(2,) + 2

3 2 Maze Example: Utility Revisited 2 3 4 s + - Intended action a: 0. T(s,a,s ) 0.8 0. Suppose s = (,) and we choose action Up. U(,) = R(,) + 0.8 x U(,2) + 0. x U(2,) + 0. x R(,) 3 2 Maze Example: Utility Revisited 2 3 4 s + - Intended action a: 0. T(s,a,s ) 0.8 0. Suppose s = (,) and we choose action Up. Move up with prob. 0.8 U(,) = R(,) + 0.8 x U(,2) + 0. x U(2,) + 0. x R(,) Move left with prob. 0. (notice the wall!) Move right with prob. 0. 3

3 2 Same with Discount 2 3 4 + - Intended action a: 0. T(s,a,s ) 0.8 0. s Suppose s = (,) and we choose action Up. U(,) = R(,) + γ (0.8 x U(,2) + 0. x U(2,) + 0. x R(,)) More General Expression If we choose action a at state s, expected future rewards are: U(s) = R(s) + γ Σ s T(s,a,s ) U(s ) 4

More General Expression Expected sum of future If we choose action a at discounted state s: rewards starting at s Reward at current state s U(s) = R(s) + γσ s T(s,a,s ) U(s ) Expected sum of future discounted rewards starting at s Probability of moving from state s to state s with action a More General Expression If we are using policy π, we choose action a=π(s) at state s, expected future rewards are: U π (s) = R(s) + γ Σ s T(s,π(s),s ) U π (s ) 5

Formal Definitions Finite set of states: S Finite set of allowed actions: A Reward function R(s) Transitions probabilities: T(s,a,s ) = P(s a,s) Utility = sum of discounted rewards: U(s 0,..) = R(s 0 ) + γr(s ) +.+ γ N R(s N ) +. Policy: π :S A Optimal policy: π*(s) = action that maximizes the expected sum of rewards from state s Markov Decision Process (MDP) Key property (Markov): P(s t+ a, s 0,..,s t ) = P(s t+ a, s t ) In words: The new state reached after applying an action depends only on the previous state and it does not depend on the previous history of the states visited in the past Markov Process 6

Markov Example 2 3 4 3 2 s 2 s + - s 0 When applying the action Right from state s 2 = (,3), the new state depends only on the previous state s 2, not the entire history {s, s 0 } T(s, a,s ) = 0.8 T(s,a 2,s) = 0.6 T(s,a 2,s) = 0.2 Graphical Notations a Prob. = 0.8 s s a Prob. = 0.4 a 2 Prob. = 0.2 a 2 Prob. = 0.6 Nodes are states Each arc corresponds to a possible transition between two states given an action Arcs are labeled by the transition probabilities 7

Example (Partial) 3 2 3 4 + Intended action a: T(s,a,s ) 0.8 2-0. 0. Up, 0.8 Up, 0. (,) Up, 0. (,2) (2,) Warning: The transitions are NOT all shown in this example! I run a company Example I can choose to either save money or spend money on advertising If I advertise, I may become famous (50% prob.) but will spend money so I may become poor If I save money, I may become rich (50% prob.), but I may also become unknown because I don t advertise What should I do? 8

S Poor & Unknown 0 ½ ½ ½ S A Rich & Unknown 0 ½ A ½ ½ ½ ½ S ½ Poor & Famous 0 S ½ A Rich & Famous 0 A Example Policies How many policies? Which one is the best policy? How to compute the optimal policy? 9

Example: Finance and Business Note: Ok, this is an overly simplified view of business models. Similar models could be used for investment decisions, etc. States: Status of the company (cash reserves, inventory, etc.) Actions: Business decisions (advertise, acquire other companies, roll out product, etc.) Uncertainty due all the external uncontrollable factors (economy, shortages, consumer confidence ) Optimal policy: The policy for making business decisions that maximizes the expected future profits Example: Robotics States are 2-D positions Actions are commanded motions (turn by x degrees, move y meters) Uncertainty comes from the fact that the mechanism is not perfect (slippage, etc.) and does not execute the commands exactly Reward when avoiding forbidden regions (for example) Optimal policy: The policy that minimizes the cost to the goal 20

Interesting example because it is impossible to store explicitly the transition probability tables (or the states, or the values U(s) ) Example: Games States: Number of white and black checkers at each location Note: Number of states is huge, on the order 0 20 states!!!! Branching factor prevents direct search Actions: Set of legal moves from any state Uncertainty comes from the roll of the dice Reward computed from number of checkers in the goal quadrant Optimal policy: The one that maximizes the probability of winning the game Example: Robotics Note : The states are continuous in this case. Although we will cover only MDPs with discrete states, the concepts can be extended to continuous spaces. Note 2: It is obviously impossible to try different policies on the system itself, for obvious reasons (it will crash to the ground on most policies!). Learning how to fly helicopters! States: possible values for the roll,pitch,yaw,elevation of the helicopter Actions: Commands to the actuators. The uncertainty comes from the fact that the actuators are imperfect and that there are unknown external effects like wind gusts Reward: High reward if it remains in stable flight (low reward if it goes unstable and crashes!) Policy: A control law that associates a command to the observed state Optimal policy: The policy that maximizes flight stability for a particular maneuver (e.g., hovering) 2

Key Result For every MDP, there exists an optimal policy There is no better option (in terms of expected sum of rewards) than to follow this policy How to compute the optimal policy? We cannot evaluate all possible policies (in real problems, the number of states is very large) Bellman s Equation If we choose an action a: U(s) = R(s) + γσ s T(s,a,s ) U(s ) 22

Bellman s Equation If we choose an action a: U(s) = R(s) + γσ s T(s,a,s ) U(s ) In particular, if we always choose the action a that maximizes future rewards (optimal policy), U(s) is the maximum U*(s) we can get over all possible choices of actions: U*(s) = R(s)+γ max a (Σ s T(s,a,s )U*(s )) Bellman s Equation U*(s)=R(s)+γ max a (Σ s T(s,a,s )U*(s )) The optimal policy (choice of a that maximizes U) is: π*(s) = argmax a (Σ s T(s,a,s ) U*(s )) 23

Why it cannot be solved directly U*(s) = R(s) + γ max a (Σ s T(s,a,s ) U*(s )) Set of S equations. Non-linear because of the max : Cannot be The optimal policy (choice of a that maximizes U) solved is: directly! π*(s) = argmax a (Σ s T(s,a,s ) U*(s )) Expected sum of rewards using policy π* The right-hand depends on the unknown. Cannot solve directly First Solution: Value Iteration Define U (s) = best value after one step U (s) = R(s) Define U 2 (s) = best possible value after two steps U 2 (s) = R(s) + γ max a (Σ s T(s,a,s ) U (s )). Define U k (s) = best possible value after k steps U k (s) = R(s) + γ max a (Σ s T(s,a,s ) U k- (s )) 24

First Solution: Value Iteration Define U (s) = best value after one step U (s) = R(s) Define U 2 (s) = best value after two steps Maximum possible expected sum of discounted rewards that I can get if I start at state s and I survive for k time steps. U 2 (s) = R(s) + γ max a (Σ s T(s,a,s ) U (s )). Define U k (s) = best value after k steps U k (s) = R(s) + γ max a (Σ s T(s,a,s ) U k- (s )) I run a company Example I can choose to either save money or spend money on advertising If I advertise, I may become famous (50% prob.) but will spend money so I may become poor If I save money, I may become rich (50% prob.), but I may also become unknown because I don t advertise What should I do? 25

S Poor & Unknown 0 ½ ½ ½ S A Rich & Unknown 0 ½ A ½ ½ ½ ½ S ½ Poor & Famous 0 S ½ A Rich & Famous 0 A Value Iteration PU PF RU RF U 0 0 0 0 U 2 0 4.5 4.5 9 U 3 2.03 8.55 6.53 25.08 U 4 4.76 2.2 8.35 28.72 U 5 7.63 5.07 20.40 3.8 U 6 0.2 7.46 22.6 33.2 U 7 2.45 9.54 24.77 35.2 U k (s) = R(s) + γ max a (Σ s T(s,a,s ) U k- (s )) 26

U(s) U k (RF) U k (RU) U k (PF) U k (PU) Iteration k Value Iteration: Facts As k increases, U k (s) converges to a value U*(s) The optimal policy is then given by: π*(s) = argmax a (Σ s T(s,a,s ) U * (s )) And U* is the utility under the optimal policy π* See convergence proof in R&N 27

U(s) Upon convergence: π*(s) = argmax a (Σ s T(s,a,s ) U*(s )) U k (RF) U k (RU) U k (PF) U k (PU) π*(pu) = A π*(pf) = S π*(ru) = S π*(rf) = S Better to always save except if poor and unknown Iteration k Maze Example 3 2 2 3 4 + - 28

Key Convergence Results Iterations γ The error on U is reduced by γ at each iteration Exponentially fast convergence Slower convergence as γ increases So far. Definition of discounted sum of rewards to measure utility Definition of Markov Decision Processes (MDP) Assumes observable states and uncertain action outcomes Optimal policy = choice of action that results in the maximum expected rewards in the future Bellman equation for general formulation of optimal policy in MDP Value iteration (dynamic programming) technique for computing the optimal policy Next: Other approaches for optimal policy computation + examples and demos. 29

Another Solution: Policy Iteration Start with a randomly chosen policy π o Iterate until convergence (π k ~ π k+ ):. Compute U k (s) for every state s using π k 2. Update the policy by choosing the best action given the utility computed at step k: π k+ (s) = argmax a (Σ s T(s,a,s ) U k (s )) The sequence of policies π o, π,, π k, converges to π* Evaluating a Policy. Compute U k (s) for every state s using π k U k (s) = R(s) + γσ s T(s, π k (s),s ) U k (s ) Linear set of equations can be solved in O( S 3 ) May be too expensive for S large, use instead simplified update: U k (s) R(s) + γσ s T(s, π k (s),s ) U k- (s ) (modified policy iteration) 30

Evaluating a Policy This is only an approximation because we should use U k here.. Compute U k (s) for every state s using π k U k (s) = R(s) + γσ s T(s, π k (s),s ) U k (s ) Linear set of equations can be solved in O( S 3 ) May be too expensive for S large, use instead simplified update: U k (s) R(s) + γσ s T(s, π k (s),s ) U k- (s ) (modified policy iteration) Comparison Value iteration: (Relatively) small number of actions Policy iteration: Large number of actions Initial guess at a good policy Combined policy/value iteration is possible Note: No need to traverse all the states in a fixed order at every iteration Random order ok Predict most useful states to visit Prioritized sweeping Choose the state with largest value to update States can be visited in any order, applying either value or policy iteration asynchronous iteration 3

Limitations We need to represent the values (and policy) for every state in principle In real problems, the number of states may be very large Leads to untractably large tables (checker-like problem with N cells and M pieces N(N-)(N-2)..(N-M) states!) Need to find a compact way of representing the states Solutions: Interpolation State s Value U Memory-based representations s Hierarchical representations s 2 s 00000 s 0000 U Function Approximation State Value States Polynomials/Splines approximation: Represent U(s) by a polynomial function that can be represented by a small number of parameters. Economic models, control Operations Research Channel routing, Radio therapy Neural Nets: Represent U(s) implicitly by a neural net (function interpolation). Elevator scheduling, Cell phones, Backgammon, etc. 32

Memory-Based Techniques States stored in memory States stored in memory U(s) =? U(s) =? Replace U(s) by U(closest neighbor to s) Replace U(s) by weighted average of U(K closest neighbor to s) Hierarchical Representations Split a state into smaller states when necessary Hierarchy of states with high-level managers directing lower-level servants Example from Dayan, Feudal learning. 33

Multiresolution: Examples from Foster & Dayan 2000 More Difficult Case Uncertainty on transition from one state to the next as before because of imperfect actuators. But, now we have also: Uncertainty on our knowledge of the state we re in because of imperfect sensors. The state is only partially observable: Partially Observable Markov Decision Process (POMDP) 34

As before: States, s Actions, a POMDP Transitions, T(s,a,s ) = P(s a,s) New: The state is not directly observable, instead: Observations, o Observation model, O(s,o) = P(o s) As before: States, s Actions, a POMDP Transitions, T(s,a,s ) = P(s a,s) New: The state is not directly observable, instead: Observations, o Probability of making observation o when at state s Observation model, O(s,o) = P(o s) 35

POMDP: What is a policy? We don t know for sure which state we re in, so it does not make sense to talk about the optimal choice of action for a state All we can define is the probability that we are in any given state: b(s) = [P(s ),,P(s N )] Policy: Choice of action for a given belief state π(b): belief state b to action a MDP with belief states instead of POMDP: states What is a policy? Unfortunately: We don t know - Requires for sure continuous which state representation we re in, so it does not make - Untractable sense to talk in general about the optimal choice of action - Approximations for a state or special cases All we can define is the probability that we are in any given state: b(s) = [P(s ),,P(s N )] Policy: Choice of action for a given belief state π(b): belief state b to action a belief state Probability that the agent is in state s k 36

Summary Definition of discounted sum of rewards to measure utility Definition of Markov Decision Processes (MDP) Assumes observable states and uncertain action outcomes Optimal policy = choice of action that results in the maximum expected rewards in the future Bellman equation for general formulation of optimal policy in MDP Value iteration technique for computing the optimal policy Policy iteration technique for computing the optimal policy MDP = generalization of the deterministic search techniques studied earlier in class POMDP = MDP + Uncertainty on the observed states 37