-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Usage]: 请问支持多机多卡吗 #9
Comments
|
感谢回复,看着是ray_npu_executor.py里worker_ip = ray.get(worker.get_node_ip.remote())开始报的
|
以前在其他三方库遇到过类似问题,没有import torch_npu导致的,你试试在报错的代码前加上import torch_npu可以修复吗 |
试了下不行,还是一样的错 |
Your current environment
我合并了npu_support和vllm_main的最新代码,单机npu上可以运行,想尝试多机,在使用ray时出现了问题
ray的环境看起没有问题,2*4npu,如下:
我的启动命令:
vllm serve Qwen2_5/Qwen2.5-72B-Instruct/ --tensor-parallel-size=4 --pipeline-parallel-size 2 --block-size=128 --trust-remote-code --uvicorn-log-level=debug
报错:please ensure that world_size (8) is less than than max local Ascend npu count (4),没有识别到ray
加了 --distributed-executor-backend=ray参数,报错:
ray.exceptions.RaySystemError: System error: Expected one of cpu, cuda, ipu, xpu, mkldnn, opengl, opencl, ideep, hip, ve, fpga, maia, xla, lazy, vulkan, mps, meta, hpu, mtia, privateuseone device type at start of device string: npu
How would you like to use vllm
希望能跑通多机
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: