Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WebNN] Remove validation for coordinate_transformation_mode #22811

Merged
merged 1 commit into from
Nov 13, 2024

Conversation

shiyi9801
Copy link
Contributor

@shiyi9801 shiyi9801 commented Nov 12, 2024

The performance cost of falling back to the CPU EP is high for several resampling nodes and causes multiple partitions in SD Turbo and VAE decoder. Since the asymmetric mode with nearest to floor and integer scales is identical to half_pixel anyway, stick with the WebNN EP.

@shiyi9801
Copy link
Contributor Author

@Honry @fdwr PTAL, thanks!

BTW, should WebNN add coordinate_transformation_mode as a parameter of Resize? There is a spec issue about it: webmachinelearning/webnn#270 @fdwr

@guschmue guschmue added the ep:WebNN WebNN execution provider label Nov 12, 2024
@fdwr
Copy link
Contributor

fdwr commented Nov 12, 2024

Remind me which model was impacted by this - what is SAM? Oh, this was for SD Turbo unet and VAE decoder model. We should update the description with more information about the motivation, e.g.:

Since performance cost of falling back to the CPU EP is high for several resampling nodes which causes multiple partitions, and since asymmetric with nearest mode to floor with integer scales is identical to half_pixel anyway, stick with the WebNN EP.

BTW, should WebNN add coordinate_transformation_mode as a parameter of Resize? There is a spec issue about it:

Ooh, nice table @Honry. We'd have to look at TFLite(EdgeRT) and CoreML first. DirectML supports basically any transformation mode with a scale factor and offsets.

Copy link
Contributor

@fdwr fdwr left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

👍

@fdwr
Copy link
Contributor

fdwr commented Nov 12, 2024

/azp run ONNX Runtime Web CI Pipeline,Windows GPU CI Pipeline,Linux Android Emulator QNN CI Pipeline

@fdwr
Copy link
Contributor

fdwr commented Nov 12, 2024

/azp run Linux CPU CI Pipeline,Linux CPU Minimal Build E2E CI Pipeline,Linux GPU CI Pipeline,Linux GPU TensorRT CI Pipeline,Linux OpenVINO CI Pipeline,Linux QNN CI Pipeline,MacOS CI Pipeline,Windows ARM64 QNN CI Pipeline,Windows CPU CI Pipeline

@fdwr
Copy link
Contributor

fdwr commented Nov 12, 2024

/azp run Windows GPU CUDA CI Pipeline,Windows GPU DML CI Pipeline,Windows GPU Doc Gen CI Pipeline

@fdwr
Copy link
Contributor

fdwr commented Nov 12, 2024

/azp run Windows GPU TensorRT CI Pipeline,onnxruntime-binary-size-checks-ci-pipeline,orttraining-linux-ci-pipeline,orttraining-linux-gpu-ci-pipeline,orttraining-ortmodule-distributed,Windows x64 QNN CI Pipeline,Big Models

Copy link

Azure Pipelines successfully started running 2 pipeline(s).

Copy link

Azure Pipelines successfully started running 3 pipeline(s).

Copy link

Azure Pipelines successfully started running 4 pipeline(s).

Copy link

Azure Pipelines successfully started running 9 pipeline(s).

@fdwr
Copy link
Contributor

fdwr commented Nov 13, 2024

(I'll update the CR description for you)

@fdwr fdwr merged commit 3adcf4d into microsoft:main Nov 13, 2024
75 checks passed
@shiyi9801
Copy link
Contributor Author

(I'll update the CR description for you)

Oh I didn't notice your comment above, thanks! :)

ishwar-raut1 pushed a commit to ishwar-raut1/onnxruntime that referenced this pull request Nov 19, 2024
…ft#22811)

The performance cost of falling back to the CPU EP is high for several
resampling nodes and causes multiple partitions in SD Turbo and VAE
decoder. Since the asymmetric mode with nearest to floor and integer
scales is identical to half_pixel anyway, stick with the WebNN EP.
guschmue pushed a commit that referenced this pull request Dec 2, 2024
The performance cost of falling back to the CPU EP is high for several
resampling nodes and causes multiple partitions in SD Turbo and VAE
decoder. Since the asymmetric mode with nearest to floor and integer
scales is identical to half_pixel anyway, stick with the WebNN EP.
ankitm3k pushed a commit to intel/onnxruntime that referenced this pull request Dec 11, 2024
…ft#22811)

The performance cost of falling back to the CPU EP is high for several
resampling nodes and causes multiple partitions in SD Turbo and VAE
decoder. Since the asymmetric mode with nearest to floor and integer
scales is identical to half_pixel anyway, stick with the WebNN EP.
ankitm3k pushed a commit to intel/onnxruntime that referenced this pull request Dec 11, 2024
…ft#22811)

The performance cost of falling back to the CPU EP is high for several
resampling nodes and causes multiple partitions in SD Turbo and VAE
decoder. Since the asymmetric mode with nearest to floor and integer
scales is identical to half_pixel anyway, stick with the WebNN EP.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ep:WebNN WebNN execution provider
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants