Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Integration with HF transformers #11

Closed
younesbelkada opened this issue Feb 7, 2024 · 5 comments · Fixed by huggingface/transformers#28928
Closed

Integration with HF transformers #11

younesbelkada opened this issue Feb 7, 2024 · 5 comments · Fixed by huggingface/transformers#28928

Comments

@younesbelkada
Copy link

Hi there,

Thanks for the great work !
We recently added HfQuantizer support in HF transformers thanks to @poedator 's work - I was wondering if it would make sense to create a new quantizer AqlmQuantizer for this method to natively support inference in HF transformers: https://huggingface.co/docs/transformers/main/en/hf_quantizer

Let me know if this makes sense!

@Vahe1994
Copy link
Owner

Vahe1994 commented Feb 8, 2024

Hello!
Thank you for bringing this up. We will take a look and will try to make this AqlmQuantizer for the method to natively support inference in HF.

@BlackSamorez
Copy link
Collaborator

I'm on it.

@younesbelkada
Copy link
Author

Fantastic work @BlackSamorez !

@BlackSamorez
Copy link
Collaborator

Now we wait for the release. After that we can merge #14 and close this issue.

@BlackSamorez
Copy link
Collaborator

Starting with transformers 4.38.0, AQLM is supported through quantizers interface.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants