Skip to content
This repository has been archived by the owner on Jan 15, 2024. It is now read-only.

For NCEDense, BCE decreases, SoftmaxCE increases #684

Answered by zjost
zjost asked this question in Q&A
Discussion options

You must be logged in to vote

I have worked on this more and discovered that if I allow training to keep running, eventually the SoftMaxCE will start to improve. It seems that the loss curve always increases for the first several epochs, but eventually starts reducing.

I'm still curious to understand why this happens, and if there are better ways to e.g. schedule the learning rate to get improved convergence. However, I don't think there's an issue with the code/implementation, so this issue can be closed. I apologize for the false alarm.

Replies: 5 comments 1 reply

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
1 reply
@tmacraft
Comment options

Answer selected by szha
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
3 participants
Converted from issue

This discussion was converted from issue #684 on August 30, 2020 19:12.