Question
Assume you have data in the form of just the following 5 complete episodes for an MRP (Markov Reward Process). Non-terminal States are labelled A
Assume you have data in the form of just the following 5 complete episodes for an MRP (Markov Reward Process). Non-terminal States are labelled A and B, the numbers in the episodes denote Rewards and all trajectories end in a terminal state T.
A 2 A 6 B 1 B 0 T
A 3 B 2 A 4 B 2 B 0 T
B 3 B 6 A 1 B 0 T
A 0 B 2 A 4 B 4 B 2 B 0 T
B 8 B 0 T
Given only this data and experience replay (repeatedly and endlessly drawing an episode at random from this pool of 5 episodes), what is the Value Function Every-Visit MonteCarlo will converge to, and what is the Value Function TD(0) (i.e., one-step TD) will converge to? Assume discount factor = 1. Note that your answer (Value Function at convergence) should be independent of step size.
Step by Step Solution
There are 3 Steps involved in it
Step: 1
Get Instant Access to Expert-Tailored Solutions
See step-by-step solutions with expert insights and AI powered tools for academic success
Step: 2
Step: 3
Ace Your Homework with AI
Get the answers you need in no time with our AI-driven, step-by-step assistance
Get Started