-
Notifications
You must be signed in to change notification settings - Fork 637
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Chatllama] support for accelerate from hf #233
Conversation
… class with comments, fix tokenizer truncate and add training scheduler
…e checkpoints of actor and reward model
…et and update model loader with a method to get the path for training statistics
…d reward models training
…ape[1] in PPO loss
"min_loss_scale": 1 | ||
}, | ||
"zero_optimization": { | ||
"stage": 2, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we should try to use stage 3 instead
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would like to, but with stage=3 is not working.
[Chatllama] fix typo of discounted_rewards in PPO loss
… into rlhf_accelerate
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
No description provided.