-
Notifications
You must be signed in to change notification settings - Fork 452
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add performance benchmark config: MPS 8da4w #8461
base: main
Are you sure you want to change the base?
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/8461
Note: Links to docs will display an error until the docs builds have been completed. ❌ 4 New Failures, 1 Cancelled JobAs of commit b00fce1 with merge base 0222074 ( NEW FAILURES - The following jobs have failed:
CANCELLED JOB - The following job was cancelled. Please retry:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
3b95dd2
to
df4b449
Compare
Added a link to the Benchmark project here: #8473 |
Looks good! Please schedule an on-demand benchmark job to test this new config on your PR before merging |
Bump up timeout threshold
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Job got cancelled due to timeout (after running 120min). Temporarily bump the threshold to 240mins to see if it can actually finish the run successfully. Debugging the slowness run can be done later.
@huydhn @yangw-dev do you have any idea why this benchmark job itself is running forever? Previous attempt was timed out after 2hours, and I can’t find any info why it’s taking so long. It looks like cancelled job won’t have any log? I temporarily bumped it up to 4hours to see if it can finish successfully, but per @manuelcandales the model shouldn’t run that slow. https://github.com/pytorch/executorch/actions/runs/13402426162/job/37438073261 |
From what I see in the previous run https://us-west-2.console.aws.amazon.com/devicefarm/home#/mobile/projects/02a2cf0f-6d9b-45ee-ba1a-a086587469e6/runs/4f1fcf14-2a4c-4364-ad4a-b9c7ecc0a783 and the current run https://us-west-2.console.aws.amazon.com/devicefarm/home#/mobile/projects/02a2cf0f-6d9b-45ee-ba1a-a086587469e6/runs/255ff74c-ad69-43b1-9014-4317e949d9ed, it's always iOS 18 that's hang. So, maybe this is something to do with the OS. In other cases, the test failed with this error
|
Also, here is the test output from the hang iOS 18
From what I see:
|
Weird. When I checked the log, I didn't find the section for each device. Now I can see it. @manuelcandales I think this pointer would help you debug the issue? #8461 (comment). Do you expect this benchmark config to run on both iOS 17 and 18? If 17 only, we should disable the run for 18. Per what Huy pointed out above, |
Adds a new performance benchmark config to keep track of performance on MPS backend when running Llama 3.2 1B inference with 8da4w quantization