You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm not sure if that's the right thing to do, but I remember reading that Pytorch and Tensorflow implement L2 regularization / weight decay differently (https://openreview.net/pdf?id=rk6qdGgCZ). So I decided to use the same version of L2 regularization as in the original Tensorflow implementation for consistency.
That indeed seems to be a mistake, thanks a lot for pointing it out! I will have a look at it on Monday.
Hi, I have read yoru code and have some questions here:
l2_reg_loss
and add it to BP loss. Why don't you use the l2_reg ofAdam
optimizer by settingweight_decay
param?I wonder why are there two dropout in one layer?
The text was updated successfully, but these errors were encountered: