-
Notifications
You must be signed in to change notification settings - Fork 6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Templates] Unify the batch inference template with an existing Data example #36401
[Templates] Unify the batch inference template with an existing Data example #36401
Conversation
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Template running as release tests: https://buildkite.com/ray-project/release-tests-pr/builds/42209 |
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Overall lgtm, but can we remove the explicit materialize call? It’s not necessary and prevents from streaming to the writes
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
Signed-off-by: Justin Yu <justinvyu@anyscale.com>
@amogkam Done. I originally had a So, the predictions are only computed 1x, rather than 2x as I originally thought. Is that correct? |
Yep that’s right it will only run 1x, barring a few extra samples |
…example (#36401) Signed-off-by: Justin Yu <justinvyu@anyscale.com>
…example (ray-project#36401) Signed-off-by: Justin Yu <justinvyu@anyscale.com>
…example (ray-project#36401) Signed-off-by: Justin Yu <justinvyu@anyscale.com> Signed-off-by: e428265 <arvind.chandramouli@lmco.com>
This PR de-duplicates the batch inference template by making it the same as the existing pytorch gpu batch inference example. There still needs to be a copy due to relative references in the docs not generating correctly when pulling the notebook code directly.
This PR also fixes some typos in the Data example and changes some code to have no warnings show up when running through the example (increasing the model + dataset size for a reasonable batch size with 4 workers + using a kwarg when initializing the resnet model with weights).
Notes
GPU utilization after (no warnings about reducing the batch size):
Related issue number
Checks
git commit -s
) in this PR.scripts/format.sh
to lint the changes in this PR.method in Tune, I've added it in
doc/source/tune/api/
under thecorresponding
.rst
file.