Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add lr_warmup_steps, max_steps values validation #1460

Merged
merged 4 commits into from
Jun 5, 2024

Conversation

shenxiangzhuang
Copy link
Contributor

Fix #1393

@Lightning-AI Lightning-AI deleted a comment Jun 5, 2024
@rasbt
Copy link
Contributor

rasbt commented Jun 5, 2024

Thanks for contributing! This looks good to me overall. Let me take care of the tests.

@rasbt
Copy link
Contributor

rasbt commented Jun 5, 2024

Actually, the tests are failing because in the README test, I am using

litgpt finetune_lora" checkpoint_dir \
--train.max_steps 1" \
...

With this PR, we will not be able to run quick test runs without also having to specify the warm-up steps via an appropriate value (the current default is 100).

I'd say maybe this should not be an error but perhaps a warning instead. What do you think @awaelchli & @Andrei-Aksionov ?

@awaelchli
Copy link
Contributor

I think the suggestion for a warning makes sense.

@rasbt
Copy link
Contributor

rasbt commented Jun 5, 2024

Thanks again for the contribution @shenxiangzhuang, these user experience improvements are well appreciated!

@rasbt rasbt merged commit 3e4fb84 into Lightning-AI:main Jun 5, 2024
9 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

LR scheduler can result in a division by 0
3 participants