-
-
Notifications
You must be signed in to change notification settings - Fork 6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
anyone test chatglm3-6b? set tensor_parallel_size=4, get wrong response #1735
Comments
hello, how to run multi card in single machine? when I set tensor_parallel_size=2,The following error occurred:
|
reinstall vllm in a new conda env? |
@gameofdimension any suggestion? (;´༎ຶД༎ຶ`) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Set tensor_parallel_size=1 or tensor_parallel_size=2. the response is OK.
my env info:
vllm==0.2.2
ray==2.8.0
transformers==4.34.0
torch==2.1.0
The text was updated successfully, but these errors were encountered: