ECS 271,
Machine Learning: Homework #5
Due: 6 May 2004
Instructor: Prof. Rao Vemuri,
rvemuri@ucdavis.edu
In the problems below, whenever necessary, use "epsilon" = 0.05 and "delta" = 0.05
1. (10 points) Let the hypothesis space be all conjunctive formulae over 10 Boolean variables x1, ..., x10. (2 points) What is the size of the hypothesis space H? (8 points) What is the estimate of m?
2. (10 + 10 points) How many training examples are needed to train a 2-input, 1-output Perceptron. Compare this estimated value with the actual value from your experience.You can look back at your own training experience with Perceptrons while doing an earlier homework problem. Or, you can run the program again.
3. (20 points) Suppose you want to train a neural net with W weights. Roughly, how many training examples are needed? Assume that we want 90% accuracy on test examples drawn from the same fixed distribution used to train the network. Is this estimate borne out in practice? You look back at your own experience from the homework exercise submitted last week. If necessary run the program again to gather any needed data.
4. (35 points) Consider a decision tree. Assume k attributes. Estimate m for PAC-larning of a decision tree.
5.
(15 points) Given
a set X, called the input space. Typically X will be a subset of Rm
for some m. You are also given a concept class C, which consists of a
family of subsets of X. Some examples of inputs and concept classes are given
below:
(a) X = R, C = infinite open
intervals, or the empty set
(b) X = R, C = R, 0, one open interval,
or two disjoint infinite open intervals
(c) X = R2, C = open
half-spaces
Show
that each of these is shattered respectively by
(d) Every 2-element set, no
3-element set
(e) Every 3-element set, no
4-element set
(f) Every non-collinear 3-element
set, no 4-element set
Support your answer by a brief but complete explanation. You can use
appropriate sketches to supplement your explanation.