Answered step by step
Verified Expert Solution
Link Copied!

Question

1 Approved Answer

Exercise 2 on the Cross-entropy cost function in NNDL 3 FromChapter 3 of NNDL, answer the second question only , which starts like: In the

Exercise 2 on the Cross-entropy cost function in NNDL 3

FromChapter 3 of NNDL, answer thesecond question only, which starts like:

"In the single-neuron discussion at the start of this section, I argued that the cross-entropy is small if(z)y

(z)yfor all training inputs..."

For this problem, you can assumeyas a single neuron (in the output layer) or a vector (i.e., the whole output layer).

  • If you assume the former, you can do a rigorous proof by using calculus and minimizing the derivative of the function. But if you are not comfortable with calculus,you can pick at least three values fory(between 0 and 1), and for each value ofy, you should compute the Cross-entropy value using thatyand several varying values ofa(e.g. 0.1, 0.2, 0.3,..., 0.9).
  • If you assume the latter, you can do a formal proof by calculus as well (although the derivative function will have several variables), but I recommend using Information Theory to prove the formula will minimize whenyandaare equal.

Step by Step Solution

There are 3 Steps involved in it

Step: 1

blur-text-image

Get Instant Access to Expert-Tailored Solutions

See step-by-step solutions with expert insights and AI powered tools for academic success

Step: 2

blur-text-image

Step: 3

blur-text-image

Ace Your Homework with AI

Get the answers you need in no time with our AI-driven, step-by-step assistance

Get Started

Recommended Textbook for

Precalculus With Limits

Authors: Ron Larson

3rd Edition

1285607163, 9781285607160

More Books

Students also viewed these Mathematics questions

Question

What functions might this behavior be serving?

Answered: 1 week ago

Question

1. What does this mean for me?

Answered: 1 week ago