-
Notifications
You must be signed in to change notification settings - Fork 27.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LlavaForConditionalGeneration logit values are batch size dependent. #29327
Comments
Also the format you are using is wrong |
#29327 (comment)
Unless you use a different transformer version, is this likely due to different system specifications? |
I am using the latest release of transformers. |
If there was a bug, it was fixe 😉 |
I am also using the latest transformers from the main branch. I cannot explain the difference between your execution (~1e-5) and mine (~1e-3). I'm unsure if this is relevant, but I use CPU since the model won't fit in my GPU without quantization. I am avoiding quantization of the base model to isolate different sources of error. |
1e-3 is already close enough. I don't know which CPU you are using but I think we can agree that the outputs are not batch dependent. |
sounds good. I am closing the issue. Thanks for the help! |
Hi @ArthurZucker,
Thanks for the response. From the comment, the possible reasons in llama models are -
I tested with float32, my observations are as follows -
I am attaching the updated script to reproduce the mismatch,
Originally posted by @ShahRutav in #29282 (comment)
The text was updated successfully, but these errors were encountered: