Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add performance benchmark config: MPS 8da4w #8429

Closed
wants to merge 2 commits into from

Conversation

manuelcandales
Copy link
Contributor

Adds a new performance benchmark config to keep track of performance on MPS backend when running Llama 3.2 1B inference with 8da4w quantization

Copy link

pytorch-bot bot commented Feb 12, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/8429

Note: Links to docs will display an error until the docs builds have been completed.

❌ 8 New Failures, 1 Cancelled Job

As of commit 7d3ca20 with merge base 931bb8b (image):

NEW FAILURES - The following jobs have failed:

CANCELLED JOB - The following job was cancelled. Please retry:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Feb 12, 2025
@guangy10
Copy link
Contributor

@manuelcandales Here is the instructions of how to trigger an on-demand benchmark job on your PR using the newly added benchmark config:

  1. On GitHub, got to "Actions" tab, and select "apple-perf" workflow
  2. Click on "Run workflow", then select your branch, then enter the HF model_id (meta-llama/Llama-3.2-1B for example) and config name llama3_mps_8da4w, and click the green "Run workflow" button (as shown in the screenshot)

Screenshot 2025-02-12 at 3 06 55 PM

@manuelcandales manuelcandales had a problem deploying to upload-benchmark-results February 12, 2025 23:09 — with GitHub Actions Failure
@guangy10
Copy link
Contributor

@manuelcandales Please note that the all infra backed by pytorch dev infra including the benchmarking infra can only run on a non-forked PR. If your PR is created on your own fork (seems like yes), you have to recreate a non-fork PR.

@manuelcandales
Copy link
Contributor Author

@guangy10 Thank you for the detailed instructions. I created a non-fork PR #8461. However, when I select my branch under Run Workflow, it doesn't show the Backend delegates field. Notice that when I select the bench-debug branch from your example, I can see it. Do you know why are these templates different from branch to branch? Am I missing something on my branch?
Screenshot 2025-02-13 at 9 54 38 AM

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. topic: not user facing
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants