CMSC 671 - Fall 2005
Homework #6
Out Tuesday 11/29/05
Due Tuesday 12/13/05
1. Decision Trees (15 pts.)
Suppose that an attribute splits the set of examples E into subsets
Ei, and that each subset has pi positive
examples and ni negative examples. Show that unless
the ratio
pi / (pi+n)
is the same for all i, the attribute has strictly positive information
gain.
2. Learning Bayes Nets (15 pts)
Russell & Norvig, Exercise 20.9. Consider an arbitrary Bayesian
network, a complete data set for that network, and the likelihood for the
data set according to the network. Give a simple proof (i.e., in words)
that the likelihood of the data cannot decrease if we add a new link to the
network and recompute the meaximum-likelihood parameter values.
3. Reinforcement Learning (40 pts)
Consider the following deterministic grid world. Allowable moves are
shown by arrows; the numbers indicate the reward for performing an action.
(Unmarked arrows mean that the action yields zero reward/penalty.)
(a) Given current values for Q shown below, show all of the changes in the Q
estimates when the agent takes the path shown by the dotted line. (Note
that the path starts in the lower left cell, and ends in the upper right
cell.) Use gamma=.5.
(b) Show all of the final optimal Q values for gamma=0.5 and for gamma=0.9.
(Mark these values on two copies of the grid.)
(c) Given your Q values above, show all of the V* values for each state, and
mark one optimal policy (action to take in each cell), for gamma=0.5 and
for gamma=0.9.
4. Nash Equilibria (15 pts)
Russell & Norvig, Exercise 17.9. Show that a dominant strategy
equilibrium is always a Nash equilibrium, but not (necessarily) vice versa.
5. Game Theory (15 pts)
Russell & Norvig, Exercise 17.11. Solve the game of three-finger
Morra.