Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The optimizer may track some values across different training routines #136

Closed
DavdGao opened this issue Jun 2, 2022 · 0 comments · Fixed by #177
Closed

The optimizer may track some values across different training routines #136

DavdGao opened this issue Jun 2, 2022 · 0 comments · Fixed by #177
Assignees

Comments

@DavdGao
Copy link
Collaborator

DavdGao commented Jun 2, 2022

  • Sine the optimizer is initialized within context, different training routines will use the same optimizer.
  • Some optimizers, like Adam, will track the past momentum.

Therefore, the optimizer may track some state variables across different training routines. Considering the initialized model of each training routine is broadcast by the server, maybe it is unnecessary or even wrong to track past variables.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant