We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Longformer takes global_attention_mask as input in the current transformers onnx export. Hence, it is currently not supported with ORTModel.
global_attention_mask
Before going forward, it could be good to benchmark longformer with transformers onnx export vs the custom https://github.com/microsoft/onnxruntime/tree/main/onnxruntime/python/tools/transformers/models/longformer
Can see how we can support, but I'm worried adding custom cases like this to ORTModel adds overhead.
The text was updated successfully, but these errors were encountered:
As an alternative, is that possible/reasonable to use other model's config to optimize longformer? such as put
{ "model_type": "bert" }
in config.json for optimize longformer? Thanks @fxmarty
Sorry, something went wrong.
No branches or pull requests
Feature request
Longformer takes
global_attention_mask
as input in the current transformers onnx export. Hence, it is currently not supported with ORTModel.Motivation
Before going forward, it could be good to benchmark longformer with transformers onnx export vs the custom https://github.com/microsoft/onnxruntime/tree/main/onnxruntime/python/tools/transformers/models/longformer
Your contribution
Can see how we can support, but I'm worried adding custom cases like this to ORTModel adds overhead.
The text was updated successfully, but these errors were encountered: