Answered step by step
Verified Expert Solution
Link Copied!

Question

1 Approved Answer

6 . Neural networks and backpropagation ( 1 0 points ) . Consider a simple two - layer network in the lecture slides. Given n

6. Neural networks and backpropagation (10 points). Consider a simple two-layer network in the lecture slides. Given
n
training data
(x
i
,y
i
)
,
i=1,...,n
, the cost function used to training the neural networks \[\ell(w,\alpha,\beta)=\sum_{i=1}^{n}\left(y^{i}-\sigma\left(w^{T} z^{i}\right)\right)^{2}\] where
\sigma (x)=1/(1+e
x
)
is the sigmoid function,
z
i
is a two-dimensional vector such that
z
1
i
=\sigma (\alpha
T
x
i
)
, and
z
2
i
=\sigma (\beta
T
x
i
)
.2(a)(5 points) Show that the gradient is given by \[\frac{\partial \ell(w,\alpha,\beta)}{\partial w}=-\sum_{i=1}^{n}2\left(y^{i}-\sigma\left(u^{i}\right)\right)\sigma\left(u^{i}\right)\left(1-\sigma\left(u^{i}\right)\right) z^{i},\] where
u
i
=w
T
z
i
. This is also known as backpropagation. (b)(5 points) Also show the gradient of
(w,\alpha ,\beta )
with respect to
\alpha
and
\beta
and write down their expression.

Step by Step Solution

There are 3 Steps involved in it

Step: 1

blur-text-image

Get Instant Access to Expert-Tailored Solutions

See step-by-step solutions with expert insights and AI powered tools for academic success

Step: 2

blur-text-image

Step: 3

blur-text-image

Ace Your Homework with AI

Get the answers you need in no time with our AI-driven, step-by-step assistance

Get Started

Recommended Textbook for

Guide To Client Server Databases

Authors: Joe Salemi

2nd Edition

1562763105, 978-1562763107

More Books

Students also viewed these Databases questions