You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There is a nvidia docker decoder and cuda function performance issue on mutiple cards we have meet. We have test on Geforce 2080 Ti and Tesla T4 cards. We used 2 cards on one server to test. At beginning, we just use nvidia docker work on cenos 7, we found that when we use 2 cards working at the same time, the performance reduced almost to half. Then we use vmware to create 2 virtual machines on one server with gpu passthrough technology, each vm bound with one card, the result show each card arrive it's best performance.The test case and snaps as follow:
The program we use nvidia Video_Codec_SDK_11.1.5 to decode , then we write cuda code transform nv12 to bgr:
2. Steps to reproduce the issue
we just run one docker cantainer on one gpu card.we put 30 channels of rtsp video to test, the performance is good. we can get valid picture from the program, the fps kept on more than 25.
nvidia-smi pmon
cuda function elapsed time
then we run one docker container with 2 gpu cards to test. we put 60 channels of rtsp video to test, the result reduce to half
nvidia-smi pmon
cuda function elapsed time
after that , we run two docker containers with 2 gpu cards, each container bound each card. we put 60 channels of rtsp video to test. we make sure that each container have 30 channels. the result is the same as on docker container with 2 cards.
so we think the container can't isolate the gpu card very well, then we use vmware to test.we created 2 virtual machines to test. we passthrough 2 cards to each machine.then run one container in each vm.The performance is well, each card is the same as the first case.
as the result, we get the issue: the docker can't isolate mulitple gpu cards, the performance reduced very quickly on mutiple gpu cards.
The text was updated successfully, but these errors were encountered:
1. Issue or feature description
There is a nvidia docker decoder and cuda function performance issue on mutiple cards we have meet. We have test on Geforce 2080 Ti and Tesla T4 cards. We used 2 cards on one server to test. At beginning, we just use nvidia docker work on cenos 7, we found that when we use 2 cards working at the same time, the performance reduced almost to half. Then we use vmware to create 2 virtual machines on one server with gpu passthrough technology, each vm bound with one card, the result show each card arrive it's best performance.The test case and snaps as follow:
The program we use nvidia Video_Codec_SDK_11.1.5 to decode , then we write cuda code transform nv12 to bgr:
2. Steps to reproduce the issue
we just run one docker cantainer on one gpu card.we put 30 channels of rtsp video to test, the performance is good. we can get valid picture from the program, the fps kept on more than 25.
nvidia-smi pmon
cuda function elapsed time
then we run one docker container with 2 gpu cards to test. we put 60 channels of rtsp video to test, the result reduce to half
nvidia-smi pmon
cuda function elapsed time
after that , we run two docker containers with 2 gpu cards, each container bound each card. we put 60 channels of rtsp video to test. we make sure that each container have 30 channels. the result is the same as on docker container with 2 cards.
so we think the container can't isolate the gpu card very well, then we use vmware to test.we created 2 virtual machines to test. we passthrough 2 cards to each machine.then run one container in each vm.The performance is well, each card is the same as the first case.
as the result, we get the issue: the docker can't isolate mulitple gpu cards, the performance reduced very quickly on mutiple gpu cards.
The text was updated successfully, but these errors were encountered: