You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, thanks for sharing the code of your great work on SphereFace Revived. I have a related question:
when using the three proposed normalization methods, why do you scaling a constant on the cross entropy loss
in sphereR_N,sphereR_H and sphereR_S.
Any help would be appreciated, thanks!
The text was updated successfully, but these errors were encountered:
thanks a lot, I found an interesting answer why scaling loss does affect the training result.
It seems that scaling the loss under SGD and no regularization equals scaling the learning rate.
Hi, thanks for sharing the code of your great work on SphereFace Revived. I have a related question:
when using the three proposed normalization methods, why do you scaling a constant on the cross entropy loss
in sphereR_N,sphereR_H and sphereR_S.
Any help would be appreciated, thanks!
The text was updated successfully, but these errors were encountered: