Backpropagation: Derivaiton of Loss Gradient
Clash Royale CLAN TAG#URR8PPP
up vote
1
down vote
favorite
In the book "Artifical Intelligence: A Modern Approach" from S. Russel there is a derivation of the gradient of the loss with respect to the weights w used for backpropagation on page 735.
I stumbled across the third row of the derivation, where $Delta_k$ is inserted.
On the left side of the equation, there is still a '2', but on the right side it is gone.
Can someone explain to me what happened to that '2'?
begineqnarray
fracpartial Loss_kpartial omega_j,k &=& -2(y_k - a_k)fracpartial a_kpartial omega_j,k = -2(y_k - a_k)fracpartial g(in_k)partial omega_j,k \
&=& -2(y_k - a_k)g'(in_k)fracpartial in_kpartial omega_j,k = -2(y_k - a_k)g'(in_k)fracpartial partial omega_j,kleft( sum _j omega_j,ka_jright) \
&=& -colorred2(y_k - a_k) g'(in_k) a_j = -a_jDelta_k
endeqnarray
where
$$
Delta_k = Err_k times g'(in_k)
$$
vector-analysis machine-learning gradient-descent neural-networks
add a comment |Â
up vote
1
down vote
favorite
In the book "Artifical Intelligence: A Modern Approach" from S. Russel there is a derivation of the gradient of the loss with respect to the weights w used for backpropagation on page 735.
I stumbled across the third row of the derivation, where $Delta_k$ is inserted.
On the left side of the equation, there is still a '2', but on the right side it is gone.
Can someone explain to me what happened to that '2'?
begineqnarray
fracpartial Loss_kpartial omega_j,k &=& -2(y_k - a_k)fracpartial a_kpartial omega_j,k = -2(y_k - a_k)fracpartial g(in_k)partial omega_j,k \
&=& -2(y_k - a_k)g'(in_k)fracpartial in_kpartial omega_j,k = -2(y_k - a_k)g'(in_k)fracpartial partial omega_j,kleft( sum _j omega_j,ka_jright) \
&=& -colorred2(y_k - a_k) g'(in_k) a_j = -a_jDelta_k
endeqnarray
where
$$
Delta_k = Err_k times g'(in_k)
$$
vector-analysis machine-learning gradient-descent neural-networks
Welcome to MSE. It is in your best interest that you type your questions (using MathJax) instead of posting links to pictures.
– José Carlos Santos
Jul 16 at 12:41
How is the quantity $Err_k$ defined?
– caverac
Jul 16 at 14:32
'Errk be the kth component of the error vector y − hw': So Errk is the same as (yk - ak)
– MJA
Jul 16 at 15:49
add a comment |Â
up vote
1
down vote
favorite
up vote
1
down vote
favorite
In the book "Artifical Intelligence: A Modern Approach" from S. Russel there is a derivation of the gradient of the loss with respect to the weights w used for backpropagation on page 735.
I stumbled across the third row of the derivation, where $Delta_k$ is inserted.
On the left side of the equation, there is still a '2', but on the right side it is gone.
Can someone explain to me what happened to that '2'?
begineqnarray
fracpartial Loss_kpartial omega_j,k &=& -2(y_k - a_k)fracpartial a_kpartial omega_j,k = -2(y_k - a_k)fracpartial g(in_k)partial omega_j,k \
&=& -2(y_k - a_k)g'(in_k)fracpartial in_kpartial omega_j,k = -2(y_k - a_k)g'(in_k)fracpartial partial omega_j,kleft( sum _j omega_j,ka_jright) \
&=& -colorred2(y_k - a_k) g'(in_k) a_j = -a_jDelta_k
endeqnarray
where
$$
Delta_k = Err_k times g'(in_k)
$$
vector-analysis machine-learning gradient-descent neural-networks
In the book "Artifical Intelligence: A Modern Approach" from S. Russel there is a derivation of the gradient of the loss with respect to the weights w used for backpropagation on page 735.
I stumbled across the third row of the derivation, where $Delta_k$ is inserted.
On the left side of the equation, there is still a '2', but on the right side it is gone.
Can someone explain to me what happened to that '2'?
begineqnarray
fracpartial Loss_kpartial omega_j,k &=& -2(y_k - a_k)fracpartial a_kpartial omega_j,k = -2(y_k - a_k)fracpartial g(in_k)partial omega_j,k \
&=& -2(y_k - a_k)g'(in_k)fracpartial in_kpartial omega_j,k = -2(y_k - a_k)g'(in_k)fracpartial partial omega_j,kleft( sum _j omega_j,ka_jright) \
&=& -colorred2(y_k - a_k) g'(in_k) a_j = -a_jDelta_k
endeqnarray
where
$$
Delta_k = Err_k times g'(in_k)
$$
vector-analysis machine-learning gradient-descent neural-networks
edited Jul 16 at 14:31
caverac
11k2927
11k2927
asked Jul 16 at 12:38
MJA
62
62
Welcome to MSE. It is in your best interest that you type your questions (using MathJax) instead of posting links to pictures.
– José Carlos Santos
Jul 16 at 12:41
How is the quantity $Err_k$ defined?
– caverac
Jul 16 at 14:32
'Errk be the kth component of the error vector y − hw': So Errk is the same as (yk - ak)
– MJA
Jul 16 at 15:49
add a comment |Â
Welcome to MSE. It is in your best interest that you type your questions (using MathJax) instead of posting links to pictures.
– José Carlos Santos
Jul 16 at 12:41
How is the quantity $Err_k$ defined?
– caverac
Jul 16 at 14:32
'Errk be the kth component of the error vector y − hw': So Errk is the same as (yk - ak)
– MJA
Jul 16 at 15:49
Welcome to MSE. It is in your best interest that you type your questions (using MathJax) instead of posting links to pictures.
– José Carlos Santos
Jul 16 at 12:41
Welcome to MSE. It is in your best interest that you type your questions (using MathJax) instead of posting links to pictures.
– José Carlos Santos
Jul 16 at 12:41
How is the quantity $Err_k$ defined?
– caverac
Jul 16 at 14:32
How is the quantity $Err_k$ defined?
– caverac
Jul 16 at 14:32
'Errk be the kth component of the error vector y − hw': So Errk is the same as (yk - ak)
– MJA
Jul 16 at 15:49
'Errk be the kth component of the error vector y − hw': So Errk is the same as (yk - ak)
– MJA
Jul 16 at 15:49
add a comment |Â
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2853367%2fbackpropagation-derivaiton-of-loss-gradient%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Welcome to MSE. It is in your best interest that you type your questions (using MathJax) instead of posting links to pictures.
– José Carlos Santos
Jul 16 at 12:41
How is the quantity $Err_k$ defined?
– caverac
Jul 16 at 14:32
'Errk be the kth component of the error vector y − hw': So Errk is the same as (yk - ak)
– MJA
Jul 16 at 15:49