You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We want to use nvidia-docker inside our slurm jobs. Inside the slurm job only the GPUs we have requested are visible when we look at nvidia-smi outputs. I was expecting if I do docker run --gpus all nvidia-smi I would also see the same number GPUs. However within the container nvidia-smi shows all gpus available on that machine. The container seems to have the correct CUDA_VISIBLE_DEVICES environment variable but not the correct permissions.
Could it be related to the fact that by default docker needs sudo and we've created a user group so that we don't need sudo?
(BTW I noticed some things are different in Fedora but we're using Ubuntu.)
ruro, akshaykulkarni07, stroykova, schneidr and rkindi