Solved–Artificial Intelligence Lab 10: Value Iteration –Solution

$30.00 $19.00

Department Of Computer Science and Engineering, IIT Palakkad Q1) [Chain MDP] The state is given by s = (x; y). There are 2 actions from each state namely A = fleft; rightg. Each action is successful with probability p, and the other action is made with probability 1 p. There are two terminal states T1…

You’ll get a: . zip file solution

 

 

Description

5/5 – (2 votes)

Department Of Computer Science and Engineering, IIT Palakkad

Q1) [Chain MDP] The state is given by s = (x; y). There are 2 actions from each state namely A = fleft; rightg. Each action is successful with probability p, and the other action is made with probability 1 p. There are two terminal states T1 and T2 (once in terminal state, the agent is stuck there forever). The reward in the L state is 1 and R state is +1, and every other state it is 0.

  1. Generate the chain environment. It should take the following inputs: length of chain and output the model i.e., the reward and transition probabilities (for a given state and action). [25 Marks]

  1. Implement the Bellman operator. It takes input as V and outputs T V . [15 Marks]

  1. Perform value iteration and output the optimal value function and optimal policy. Start with

various values V0 and plot jjVt V jj1. [10 Marks]

Q2) [Grid MDP] The state is given by s = (x; y). There are 4 actions from each state namely A = fup; down; left; rightg. Each action is successful with probability p, and with probability 1 3 p other 3 actions are chosen.

  1. Generate a grid environment. It should take the following inputs: x-size, y-size, goal state, blocked states, and outputs the model, i.e., the reward and transition probabilities (for a given state and action). [20 Marks]

  1. Perform value iteration and output the optimal value function and optimal policy. [10 Marks]

Q3) [Mountain Car: Deterministic Control] There is an under-powered car stuck in the bottom of a 1-dim valley. It needs to find its way to the top. The car has three actions namely A=-1,0,+1 which means accelerate backward, no acceleration and accelerate forward respectively. The ranges for position and velocity are [-1.2,0.5] and [-0.07,0.07] respectively. The car is needs to reach the top on the right, i.e., position of 0.5. The dynamics is according to the equations:

vt+1

= vt + 0:001at

0:0025cos(3pt)

pt+1

= pt + vt

(1)

  1. Perform value iteration and output the optimal value function and optimal policy. [20 Marks] (Hint: Discretise the state space into 100 100 grid (i.e., divide the position and velocity co-ordinates into 100 intervals each.)

2