Answered step by step
Verified Expert Solution
Link Copied!

Question

1 Approved Answer

QUESTION FIN 34 Martin prepared his final accounts and calculated that his net profit for the year was 7,875. This was before the following errors

QUESTION FIN 34

Martin prepared his final accounts and calculated that his net profit for the year was 7,875. This was before the following errors were discovered: (i) Rent receivable of 520 was outstanding. (ii) Sales of 928 had not been recorded. (iii) A provision for bad debts of 700 should have been created. (iv) Depreciation of 800 had not been provided for. (v) An invoice for telephone charges of 120 had not been paid. (vi) Sales returns of 322 had not been entered. (vii) Rent of 1,000 shown in the profit and loss account relates to the following year. (viii) Closing stock had been overvalued by 648. Show the corrected net profit for the year.

Compute the Gini index for the entire data set of Table 10.1, with respect to the two

classes. Compute the Gini index for the portion of the data set with age at least 50.

2. Repeat the computation of the previous exercise with the use of the entropy criterion.

3. Show how to construct a (possibly overfitting) rule-based classifier that always exhibits

100 % accuracy on the training data. Assume that the feature variables of no two

training instances are identical.

4. Design a univariate decision tree with a soft maximum-margin split criterion borrowed

from SVMs. Suppose that this decision tree is generalized to the multivariate case.

How does the resulting decision boundary compare with SVMs? Which classifier can

handle a larger variety of data sets more accurately?

5. Discuss the advantages of a rule-based classifier over a decision tree.

6. Show that an SVM is a special case of a rule-based classifier. Design a rule-based

classifier that uses SVMs to create an ordered list of rules.

7. Implement an associative classifier in which only maximal patterns are used for classification, and the majority consequent label of rules fired, is reported as the label of

the test instance.

8. Suppose that you had d-dimensional numeric training data, in which it was known that

the probability density of d-dimensional data instance X in each class i is proportiona

Explain the relationship of mutual exclusiveness and exhaustiveness of a rule set, to

the need to order the rule set, or the need to set a class as the default class.

10. Consider the rules Age > 40 ? Donor and Age ? 50 ? Donor. Are these two rules

mutually exclusive? Are these two rules exhaustive?

11. For the example of Table 10.1, determine the prior probability of each class. Determine

the conditional probability of each class for cases where the Age is at least 50.

12. Implement the naive Bayes classifier.

13. For the example of Table 10.1, provide a single linear separating hyperplane. Is this

separating hyperplane unique?

14. Consider a data set containing four points located at the corners of the square. The

two points on one diagonal belong to one class, and the two points on the other

diagonal belong to the other class. Is this data set linearly separable? Provide a proof.

15. Provide a systematic way to determine whether two classes in a labeled data set are

linearly separable.

16. For the soft SVM formulation with hinge loss, show that:

(a) The weight vector is given by the same relationship W = n

i=1 ?iyiXi, as for

hard SVMs.

(b) The condition n

i=1 ?iyi = 0 holds as in hard SVMs.

(c) The Lagrangian multipliers satisfy ?i ? C.

(d) The Lagrangian dual is identical to that of hard SVMs.

17. Show that it is possible to omit the bias parameter b from the decision boundary of

SVMs by suitably preprocessing the data set. In other words, the decision boundary is

now W X = 0. What is the impact of eliminating the bias parameter on the gradient

ascent approach for Lagrangian dual optimization in SVMs?

18. Show that an nd data set can be mean-centered by premultiplying it with the nn

matrix (I ? U), where U is a unit matrix of all ones. Show that an n n kernel

matrix K can be adjusted for mean centering of the data in the transformed space by

adjusting it to K = (I ? U)K(I ? U).

19. Consider two classifiers A and B. On one data set, a 10-fold cross validation shows

that classifier A is better than B by 3 %, with a standard deviation of 7 % over 100

different folds. On the other data set, classifier B is better than classifier A by 1 %,

with a standard deviation of 0.1 % over 100 different folds. Which classifier would you

prefer on the basis of this evidence, and why?

20. Provide a nonlinear transformation which would make the data set of Exercise 14

linearly separable.

21. Let Sw and Sb be defined according to Sect. 10.2.1.3 for the binary class problem.

Let the fractional presence of the two classes be p0 and p1, respectively. Show that

Sw + p0p1Sb is equal to the covariance matrix of the data se

image text in transcribedimage text in transcribedimage text in transcribed

Consider the stochastic trend model Yt=Bo+Bit+ne with nt defined as a simple random walk n =-1+et (et is a stationary white noise process). Show that this model can be written as I Y = B1+ Y-1 + 3. The stochastic trend model (random walk with drift) is given by with yt = Bo+ Bit+m defined as a simple random walk n =-1+e. The deterministic trend model is given by where et is a stationary white noise process. Explain the difference between these two types of trend model. Will they give similar forecasts? Why? Which one will generate forecasts with more volatility?

Step by Step Solution

There are 3 Steps involved in it

Step: 1

blur-text-image

Get Instant Access to Expert-Tailored Solutions

See step-by-step solutions with expert insights and AI powered tools for academic success

Step: 2

blur-text-image

Step: 3

blur-text-image

Ace Your Homework with AI

Get the answers you need in no time with our AI-driven, step-by-step assistance

Get Started

Recommended Textbook for

Intermediate Financial Management

Authors: Brigham, Daves

10th Edition

978-1439051764, 1111783659, 9780324594690, 1439051763, 9781111783655, 324594690, 978-1111021573

More Books

Students also viewed these Finance questions

Question

Find the probabilities. P (Z Answered: 1 week ago

Answered: 1 week ago

Question

Gambling by student and professional athletes

Answered: 1 week ago