Back propagation equation proof
up vote
0
down vote
favorite
I am trying to prove this equation (from the backpropagation equations in AI).
$$frac{partial C}{partial b_j^l} = delta_j^l$$
C is the cost function: $C = frac{1}{2}||y - a^L||^2$
Where the output of layer l and neuron j is express like so $a^l_j=σ(∑_kw^l_{jk}a^{l−1}_k+b^l_j)$
I am suppose to use this assertion to do the demonstration: $delta_j^L = frac{partial C}{partial z_j^L}$
So far, here is what I have tried:
$$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial z_j^L} frac{partial b_j^l}{partial b_j^L} $$ (I am using the chain rule to have a sum)
<=> $$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial b_j^l} frac{partial b_j^l}{partial z_j^L} $$
So I guess I have to prove, that $frac{partial b_j^l}{partial z_j^L}$ equals 1.
But I don't have any ideas how to prove it.
Thanks for your help
N.B
I am following this course => http://neuralnetworksanddeeplearning.com/chap2.html where the first two equations of the BackPropagation equations are already proved, and the 2 others should be proved the same way (using the chain rule)
artificial-intelligence
add a comment |
up vote
0
down vote
favorite
I am trying to prove this equation (from the backpropagation equations in AI).
$$frac{partial C}{partial b_j^l} = delta_j^l$$
C is the cost function: $C = frac{1}{2}||y - a^L||^2$
Where the output of layer l and neuron j is express like so $a^l_j=σ(∑_kw^l_{jk}a^{l−1}_k+b^l_j)$
I am suppose to use this assertion to do the demonstration: $delta_j^L = frac{partial C}{partial z_j^L}$
So far, here is what I have tried:
$$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial z_j^L} frac{partial b_j^l}{partial b_j^L} $$ (I am using the chain rule to have a sum)
<=> $$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial b_j^l} frac{partial b_j^l}{partial z_j^L} $$
So I guess I have to prove, that $frac{partial b_j^l}{partial z_j^L}$ equals 1.
But I don't have any ideas how to prove it.
Thanks for your help
N.B
I am following this course => http://neuralnetworksanddeeplearning.com/chap2.html where the first two equations of the BackPropagation equations are already proved, and the 2 others should be proved the same way (using the chain rule)
artificial-intelligence
add a comment |
up vote
0
down vote
favorite
up vote
0
down vote
favorite
I am trying to prove this equation (from the backpropagation equations in AI).
$$frac{partial C}{partial b_j^l} = delta_j^l$$
C is the cost function: $C = frac{1}{2}||y - a^L||^2$
Where the output of layer l and neuron j is express like so $a^l_j=σ(∑_kw^l_{jk}a^{l−1}_k+b^l_j)$
I am suppose to use this assertion to do the demonstration: $delta_j^L = frac{partial C}{partial z_j^L}$
So far, here is what I have tried:
$$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial z_j^L} frac{partial b_j^l}{partial b_j^L} $$ (I am using the chain rule to have a sum)
<=> $$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial b_j^l} frac{partial b_j^l}{partial z_j^L} $$
So I guess I have to prove, that $frac{partial b_j^l}{partial z_j^L}$ equals 1.
But I don't have any ideas how to prove it.
Thanks for your help
N.B
I am following this course => http://neuralnetworksanddeeplearning.com/chap2.html where the first two equations of the BackPropagation equations are already proved, and the 2 others should be proved the same way (using the chain rule)
artificial-intelligence
I am trying to prove this equation (from the backpropagation equations in AI).
$$frac{partial C}{partial b_j^l} = delta_j^l$$
C is the cost function: $C = frac{1}{2}||y - a^L||^2$
Where the output of layer l and neuron j is express like so $a^l_j=σ(∑_kw^l_{jk}a^{l−1}_k+b^l_j)$
I am suppose to use this assertion to do the demonstration: $delta_j^L = frac{partial C}{partial z_j^L}$
So far, here is what I have tried:
$$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial z_j^L} frac{partial b_j^l}{partial b_j^L} $$ (I am using the chain rule to have a sum)
<=> $$frac{partial C}{partial z_j^L} = sum_k frac{partial C}{partial b_j^l} frac{partial b_j^l}{partial z_j^L} $$
So I guess I have to prove, that $frac{partial b_j^l}{partial z_j^L}$ equals 1.
But I don't have any ideas how to prove it.
Thanks for your help
N.B
I am following this course => http://neuralnetworksanddeeplearning.com/chap2.html where the first two equations of the BackPropagation equations are already proved, and the 2 others should be proved the same way (using the chain rule)
artificial-intelligence
artificial-intelligence
asked Nov 18 at 13:41
Unepierre
11
11
add a comment |
add a comment |
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3003546%2fback-propagation-equation-proof%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown