7.9 AdaBoost example. In this exercise we consider a concrete example that consists of eight training points
Question:
7.9 AdaBoost example.
In this exercise we consider a concrete example that consists of eight training points and eight weak classiers.
(a) Dene an mn matrix M where Mij = yihj(xi), i.e., Mij = +1 if training example i is classied correctly by weak classier hj , and ????1 otherwise. Let dt;t 2 Rn, kdtk1 = 1 and dt;i (respectively t;i) equal the ith component of dt (respectively t). Now, consider AdaBoost as described in gure 7.7 and dene M as below with eight training points and eight weak classiers.
M = 0 BBBBBBBBBBBBBB@
????1 1 1 1 1 ????1 ????1 1 ????1 1 1 ????1 ????1 1 1 1 1 ????1 1 1 1 ????1 1 1 1 ????1 1 1 ????1 1 1 1 1 ????1 1 ????1 1 1 1 ????1 1 1 ????1 1 1 1 1 ????1 1 1 ????1 1 1 1 ????1 1 1 1 1 1 ????1 ????1 1 ????1 1 CCCCCCCCCCCCCCA Assume that we start with the following initial distribution over the datapoints:
d1 =
3 ???? p5 8 ;
3 ???? p5 8 ;
1 6 ;
1 6 ;
1 6 ;
p5 ???? 1 8 ;
p5 ???? 1 8 ; 0
>
Compute the rst few steps of the AdaBoost algorithm using M, d1, and tmax = 7. What weak classier is picked at each round of boosting? Do you notice any pattern?
(b) What is the L1 norm margin produced by AdaBoost for this example?
(c) Instead of using AdaBoost, imagine we combined our classiers using the following coecients: [2; 3; 4; 1; 2; 2; 1; 1] 1 16 . What is the margin in this case? Does AdaBoost maximize the margin?
Step by Step Answer:
Foundations Of Machine Learning
ISBN: 9780262351362
2nd Edition
Authors: Mehryar Mohri, Afshin Rostamizadeh