-
Notifications
You must be signed in to change notification settings - Fork 27.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support ONNX export for causal LM sequence classifiers #27450
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for the fix @dwyatte, LGTM! Feel free to open a PR in optimum as well to re-enable the export of those models for text-classification.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice followup to #24979. Not really sure why I did not use int
at the time, but I ran the slow tests and this seems to be alright! Thanks 🤗
The failing test might just need a rebase to main otherwise I'll skip it on main and work on a fix |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for fixing!
14c5099
to
8076250
Compare
@ArthurZucker I rebased in 8076250 but looks like something is still up with CI. Perhaps different tests get selected based on files changed or between PRs/main
|
@dwyatte Yes, the test fetcher selects a subset of the tests to run based on the files that are touched. In this case, the failing tests (I believe) are unreleated to your PR. The tests involving safetensors have had a patch pushed on main. Could you rebase on main to include these in the test runners? |
8076250
to
d8ab2c9
Compare
@amyeroberts I think there are some other problems on
|
Sorry both for the delays, I'll skip these 3 tests as well. cc @gante I'll look into the test_assisted_decoding_sample. |
Hi, @ArthurZucker , regarding the failures caused by |
just merged #27508 which should skip it for all models |
d8ab2c9
to
fef0c8b
Compare
Thanks @ArthurZucker that took care of the remaining failures. This is ready to merge |
…7450) support onnx for causal lm sequence classification
What does this PR do?
Partial fix for huggingface/optimum#1527 in
optimum
when exporting causal LMs with sequence classification support to ONNXONNX's argmax operator does not support int64, but that should not be needed here since these are just boolean tensors
Before submitting
Pull Request section?
to it if that's the case.
documentation guidelines, and
here are tips on formatting docstrings.
Who can review?
@ArthurZucker and @younesbelkada (CC @fxmarty)