-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: KeyError: 'type'. when inferencing Llama 3.2 3B Instruct #8855
Comments
It looks like vLLM wasn't installed properly. How did you install it? |
Nvm, it's a bug on our end (fixed by #8894). Is your |
@WoosukKwon I think Colab uses TPU? In that case you may have a better idea |
Weird, I was sure I had responded here and closed the issue. Anyway:
Also, I was able to get running with this:
then:
|
Your current environment
The output of `python collect_env.py`
Model Input Dumps
No response
🐛 Describe the bug
Error:
Reproduction (T4 on Colab):
There seem to be quite a few different errors with different models, like
google/gemma-2-2b-it
which gives:or
microsoft/Phi-3-mini-128k-instruct
gives:Before submitting a new issue...
The text was updated successfully, but these errors were encountered: