Run the AIPython (aipython.org) neural network code or an other learner on the Mail reading data of
Question:
Run the AIPython (aipython.org) neural network code or an other learner on the “Mail reading” data of Figure 7.1 (page 268) with a single hidden layer with two hidden units.
(a) Suppose that you decide to use any predicted value from the neural network greater than 0.5 as true, and any value less than 0.5 as false. How many examples are misclassified initially? How many examples are misclassified after 40 iterations? How many examples are misclassified after 80 iterations?
(b) Try the same example and the same initial values, with different step sizes for the gradient descent. Try at least η = 0.1, η = 1.0, and η = 5.0. Comment on the relationship between step size and convergence.
(c) Given the final parameter values you found, give a logical formula for what each of the units is computing. [Hint: As a brute-force method, for each of the units, build the truth tables for the input values and determine the output for each combination, then simplify the resulting formula.] Is it always possible to find such a formula?
(d) All of the parameters were set to different initial values. What happens if the parameter values are all set to the same (random) value? Test it out for this example, and hypothesize what occurs in general.
(e) For the neural network algorithm, comment on the following stopping criteria.
(i) Learn for a limited number of iterations, where the limit is set initially.
(ii) Stop when the squared error is less than some threshold close to zero.
(iii) Stop when the derivatives all become within some of zero.
(iv) Split the data into training data and validation data, train on the training data and stop when the error on the validation data increases.
Which would you expect to better handle overfitting? Which criteria guarantee the gradient descent will stop? Which criteria would guarantee that, if it stops, the network can be used to predict the test data accurately?
Step by Step Answer:
Artificial Intelligence: Foundations Of Computational Agents
ISBN: 9781009258197
3rd Edition
Authors: David L. Poole , Alan K. Mackworth