Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Cluster] ray running inside docker on a cloud VM losing GPU access after few hours #46552

Closed
lobo1586 opened this issue Jul 10, 2024 · 2 comments
Labels
bug Something that is supposed to be working; but isn't core Issues that should be addressed in Ray Core triage Needs triage (eg: priority, bug/not-bug, and owning component)

Comments

@lobo1586
Copy link

What happened + What you expected to happen

I believe i am currently affected by NVIDIA/nvidia-docker#1671

We deployed ray serve to a cloud compute VM using cluster_config.yaml, after running couple hours, the docker suddenly lost access to the host GPU.

This post seems to published some workaround: NVIDIA/nvidia-docker#1730

However i am not sure how can i integrate any of those solutions into the ray up cluster_config.yaml process

Versions / Dependencies

our container base image is from rayproject/ray-ml:2.24.0-py310-gpu
our GCP boot disk source image is from: projects/ml-images/global/images/c0-deeplearning-common-cu118-v20240613-debian-11-py310

Reproduction script

see linked nvidia issue

Issue Severity

High: It blocks me from completing my task.

@lobo1586 lobo1586 added bug Something that is supposed to be working; but isn't triage Needs triage (eg: priority, bug/not-bug, and owning component) labels Jul 10, 2024
@lobo1586
Copy link
Author

ok, looks like it can be fixed using nvidia's posted workaround mentioned in NVIDIA/nvidia-docker#1730, however this would require a one time fix effort, but not sure if feasible it is when there are multiple worker node with GPU

@anyscalesam anyscalesam added the core Issues that should be addressed in Ray Core label Jul 15, 2024
@rynewang
Copy link
Contributor

If I read this correctly, the workaround involves running a script on each node. Can you try to put the command here in the cluster_config.yaml? https://docs.ray.io/en/latest/cluster/vms/references/ray-cluster-configuration.html#initialization-commands

Generally speaking we expect NVIDIA to fix the issue for good, such that we don't need extra commands to fix.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something that is supposed to be working; but isn't core Issues that should be addressed in Ray Core triage Needs triage (eg: priority, bug/not-bug, and owning component)
Projects
None yet
Development

No branches or pull requests

3 participants