Felix Abecassis edited this page Nov 14, 2017 · 13 revisions

This section was kept for archiving reasons, it doesn't necessarily reflect nvidia-docker 2.0.

In this section, we will dive into the internals of nvidia-docker 1.0 to explain the problems we are solving, and how. We will also present other options for solving these problems when possible.

Enabling GPU support in your application

This section should give you enough knowledge to be able to integrate NVIDIA GPU support for your own container deployment application, for instance if you don’t want to rely on the nvidia-docker 1.0 CLI wrapper. If this is your goal, be aware that it's actually already possible to avoid relying on the nvidia-docker wrapper. Using the REST API of the nvidia-docker-plugin daemon you can launch GPU containers using only docker:

$ curl -s http://localhost:3476/docker/cli
--device=/dev/nvidiactl --device=/dev/nvidia-uvm --device=/dev/nvidia3 --device=/dev/nvidia2 --device=/dev/nvidia1 --device=/dev/nvidia0 --volume-driver=nvidia-docker --volume=nvidia_driver_361.48:/usr/local/nvidia:ro
$ docker run -ti --rm `curl -s http://localhost:3476/docker/cli` nvidia/cuda nvidia-smi

However, this method will not be able to check if the image you are using is compatible with your installed driver, as described in the image inspection subsection. Remote deployment will also be more challenging.

Another option is to leverage our nvidia package written in the Go programming language. Go is currently the language of choice for containers, many projects should be able to interface with our package seamlessly.

In-depth topics

In this section we provide details on the three major services provided by nvidia-docker:

You can’t perform that action at this time.
You signed in with another tab or window. Reload to refresh your session. You signed out in another tab or window. Reload to refresh your session.
Press h to open a hovercard with more details.