Skip to content

docker sees all GPUs despite access configured via Slurm (cgroups) #211

@mil-ad

Description

@mil-ad

We want to use nvidia-docker inside our slurm jobs. Inside the slurm job only the GPUs we have requested are visible when we look at nvidia-smi outputs. I was expecting if I do docker run --gpus all nvidia-smi I would also see the same number GPUs. However within the container nvidia-smi shows all gpus available on that machine. The container seems to have the correct CUDA_VISIBLE_DEVICES environment variable but not the correct permissions.

Could it be related to the fact that by default docker needs sudo and we've created a user group so that we don't need sudo?

(BTW I noticed some things are different in Fedora but we're using Ubuntu.)

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions