Suppose you are running a learning experiment on a new algorithm. You have a data set consisting
Question:
Suppose you are running a learning experiment on a new algorithm. You have a data set consisting of 2 examples of each of two classes. Yon plan to use leave-one-nut cross-validation. As a baseline, you run your experimental setup on a simple majority classifier. (A majority classifier is given a set of training data and then always outputs the class that is in the majority in the training set, regardless of the input.) You expect the majority classifier to score about 50% on leave-one-out cross-validation, but to your surprise, it scores zero. Can you explain why?
Fantastic news! We've Found the answer you've been seeking!
Step by Step Answer:
Related Book For
Artificial Intelligence A Modern Approach
ISBN: 978-0137903955
2nd Edition
Authors: Stuart J. Russell and Peter Norvig
Question Posted: