Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Setting FocalLossGamma = 2 causes weight corruption in the beginning of the seq2seq model training #73

Open
zsogitbe opened this issue Oct 7, 2023 · 0 comments

Comments

@zsogitbe
Copy link
Contributor

zsogitbe commented Oct 7, 2023

Description of the bug:
Setting FocalLossGamma = 2 in a sequence to sequence model training causes weight corruption in the beginning of the training and the training stops (weight corruption checking feature added recently). It is a non-random error which is always causing weight corruption in the beginning of the training. If FocalLossGamma is set to 0 the error does not happen.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant