Answered step by step
Verified Expert Solution
Link Copied!

Question

1 Approved Answer

Consider the training data given in Table 2 for classification, where the two classes of interest are and + We want to apply binary decision

image text in transcribed

Consider the training data given in Table 2 for classification, where the two classes of interest are and + We want to apply binary decision trees as our chosen algorithm for classifying this data. YES Y-4 Y=3 Y-21 Y-1 X=1 X2 Table 2 a. [6 points) Find a decision tree which uses minimum number of splits (decision boundaries at internal nodes) to perfectly classify each training data instance of Table 2. Hint: The minimum number of splits that you need to create a perfect classifier is 6. You are not required to compute the Information Gain at each split for constructing the decision tree, but to arrive at your solution by visually inspecting the data. b. 15 points Use a pessimistic estimate of the generalization error to prune this tree using sub-tree replacement post-pruning method. Use 122 as the cost of adding a leaf node while calculating the pessimistic estimate. In case there is a tie in determining the majority class of a leaf node, use as the default majority class. c. 13 points] Compare the nature of the original tree (with perfect classification on training data) with the pruned true by visual inspection. Would you use the original tree or the pruned tree for classifying any future instance of the data? State the name of the phenomena being explored in this question. Consider the training data given in Table 2 for classification, where the two classes of interest are and + We want to apply binary decision trees as our chosen algorithm for classifying this data. YES Y-4 Y=3 Y-21 Y-1 X=1 X2 Table 2 a. [6 points) Find a decision tree which uses minimum number of splits (decision boundaries at internal nodes) to perfectly classify each training data instance of Table 2. Hint: The minimum number of splits that you need to create a perfect classifier is 6. You are not required to compute the Information Gain at each split for constructing the decision tree, but to arrive at your solution by visually inspecting the data. b. 15 points Use a pessimistic estimate of the generalization error to prune this tree using sub-tree replacement post-pruning method. Use 122 as the cost of adding a leaf node while calculating the pessimistic estimate. In case there is a tie in determining the majority class of a leaf node, use as the default majority class. c. 13 points] Compare the nature of the original tree (with perfect classification on training data) with the pruned true by visual inspection. Would you use the original tree or the pruned tree for classifying any future instance of the data? State the name of the phenomena being explored in this

Step by Step Solution

There are 3 Steps involved in it

Step: 1

blur-text-image

Get Instant Access to Expert-Tailored Solutions

See step-by-step solutions with expert insights and AI powered tools for academic success

Step: 2

blur-text-image_2

Step: 3

blur-text-image_3

Ace Your Homework with AI

Get the answers you need in no time with our AI-driven, step-by-step assistance

Get Started

Recommended Textbook for

Filing And Computer Database Projects

Authors: Jeffrey Stewart

2nd Edition

007822781X, 9780078227813

More Books

Students also viewed these Databases questions

Question

6. Explain the strengths of a dialectical approach.

Answered: 1 week ago