cs344 : introduction to artificial intelligence pushpak bhattacharyya cse dept., iit bombay lecture...

Post on 14-Dec-2015

219 Views

Category:

Documents

3 Downloads

Preview:

Click to see full reader

TRANSCRIPT

CS344 : Introduction to Artificial Intelligence

Pushpak BhattacharyyaCSE Dept., IIT Bombay

Lecture 19- Probabilistic Planning

Example : Blocks World•STRIPS : A planning system – Has rules with precondition deletion list and addition list

on(B, table)on(A, table) on(C, A)hand emptyclear(C)clear(B)

on(C, table)on(B, C) on(A, B)hand emptyclear(A)

AC

A

CBB

START GOAL

Robot hand

Robot hand

Rules•R1 : pickup(x)

Precondition & Deletion List : handempty, on(x,table), clear(x)

Add List : holding(x)

•R2 : putdown(x)Precondition & Deletion List : holding(x)Add List : handempty, on(x,table), clear(x)

Rules•R3 : stack(x,y)

Precondition & Deletion List :holding(x), clear(y) Add List : on(x,y), clear(x), handempty

•R4 : unstack(x,y)Precondition & Deletion List : on(x,y),

clear(x),handemptyAdd List : holding(x), clear(y)

Plan for the block world problem

• For the given problem, Start Goal can be achieved by the following sequence :1. Unstack(C,A)2. Putdown(C)3. Pickup(B)4. Stack(B,C)5. Pickup(A)6. Stack(A,B)

• Execution of a plan: achieved through a data structure called Triangular Table.

Why Probability?

(discussion based on the book “Automated Planning” by Dana Nau)

Motivation In many situations, actions may have

more than one possible outcome Action failures

e.g., gripper drops its load Exogenous events

e.g., road closed Would like to be able to plan in such situations One approach: Markov Decision Processes

acb

Graspblock c

a

c

b

Intendedoutcome

a b c

Unintendedoutcome

Stochastic Systems

Stochastic system: a triple = (S, A, P) S = finite set of states A = finite set of actions Pa (s | s) = probability of going to s

if we execute a in s s S Pa (s | s) = 1

Robot r1 startsat location l1 State s1 in

the diagram Objective is to

get r1 to location l4 State s4 in

the diagram

Goal

Start

Example

No classical plan (sequence of actions) can be a solution, because we can’t guarantee we’ll be in a state where the next action is applicable

e.g., π =

move(r1,l1,l2), move(r1,l2,l3), move(r1,l3,l4)

Goal

Start

Example

Goal

π1 = {(s1, move(r1,l1,l2)), (s2, move(r1,l2,l3)), (s3, move(r1,l3,l4)), (s4, wait), (s5, wait)}

π2 = {(s1, move(r1,l1,l2)), (s2, move(r1,l2,l3)), (s3, move(r1,l3,l4)), (s4, wait), (s5, move(r1,l5,l4))}

π3 = {(s1, move(r1,l1,l4)), (s2, move(r1,l2,l1)), (s3, move(r1,l3,l4)), (s4, wait), (s5, move(r1,l5,l4)}

Policy: a function that maps states into actions Write it as a set of state-action pairs

Policies

Start

For every state s,there will be aprobability P(s)that the system beginsin the state s

Goal

Start

Initial States

Goal

Histories

Start

History: sequenceof system states

h = s0, s1, s2, s3, s4, …

h0 = s1, s3, s1, s3, s1, …

h1 = s1, s2, s3, s4, s4, …

h2 = s1, s2, s5, s5, s5, …

h3 = s1, s2, s5, s4, s4, …

h4 = s1, s4, s4, s4, s4, …

h5 = s1, s1, s4, s4, s4, …

h6 = s1, s1, s1, s4, s4, …

h7 = s1, s1, s1, s1, s1, … Each policy induces a probability

distribution over histories If h = s0, s1, … then P(h

| π) = P(s0) i ≥ 0 Pπ(Si) (si+1 | si)

mo

ve(r1

,l2,l1

)

Hidden Markov Models

Hidden Markov Model Set of states : S where |S|=N Output Alphabet : V Transition Probabilities : A = {aij} Emission Probabilities : B = {bj(ok)} Initial State Probabilities : π

),,( BA

Three Basic Problems of HMM

1. Given Observation Sequence O ={o1… oT} Efficiently estimate P(O|λ)

2. Given Observation Sequence O ={o1… oT} Get best Q ={q1… qT} i.e.

Maximize P(Q|O, λ)

3. How to adjust to best maximize Re-estimate λ

),,( BA)|( OP

Solutions

Problem 1: Likelihood of a sequence Forward Procedure Backward Procedure

Problem 2: Best state sequence Viterbi Algorithm

Problem 3: Re-estimation Baum-Welch ( Forward-Backward

Algorithm )

Problem 2

Given Observation Sequence O ={o1… oT}

Get “best” Q ={q1… qT} i.e.

Solution :1. Best state individually likely at a position

i2. Best state given all the previously

observed states and observations Viterbi Algorithm

Example

Output observed – aabb What state seq. is most probable? Since state

seq. cannot be predicted with certainty, the machine is given qualification “hidden”.

Note: ∑ P(outlinks) = 1 for all states

Probabilities for different possible seq

1

1,21,10.4

1,1,10.16 1,1,20.06 1,2,1 0.0375 1,2,20.0225

1,1,1,1

0.016

1,1,1,2

0.056

...and so on

1,1,2,1

0.018

1,1,2,2

0.018

0.15

IfP(si|si-1, si-2) (order 2 HMM)

then the Markovian assumption will take effect only after two levels.(generalizing for n-order… after n levels)

Viterbi for higher order HMM

Viterbi Algorithm• Define such that,

i.e. the sequence which has the best joint probability so far.

• By induction, we have,

Viterbi Algorithm

Viterbi Algorithm

top related