Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

AdamAsync-Optimizer.md中提到的adam更新公式切换到原始版本,解决nan问题,请问nan问题是如何引起的 #797

Open
hyspe3800 opened this issue Apr 6, 2023 · 1 comment

Comments

@hyspe3800
Copy link

文档:https://deeprec.readthedocs.io/zh/latest/AdamAsync-Optimizer.html

@bitzpy
Copy link

bitzpy commented May 15, 2023

我觉得可能是以下原因,下面两张图分别是adam 和 adamasync的代码。adam的代码我写注释的地方是那行代码的等价版本,他异步训练的时候-v +v两个v的取值不一样,可能导致v变成负数,开方出现nan。
image
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants