GPU memory is not released after prediction #6977
Replies: 3 comments
-
hi, there are several reasons that could cause the problem. You can check according to the following steps:
|
Beta Was this translation helpful? Give feedback.
-
@Evezerest Thanks so much for your reply and your suggestions. Just provide more details here. We are using a NVIDIA T4 Tensor Core GPU with 16GB memory. We allow maximum 8 PaddleOCR workers running during the peak time.
So, based on our observation, we suspect there is a memory leak inside Paddle when using ResNet18. Circling back to your suggestions:
|
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
Hello, I just realized that the GPU memory seems not to be released each time after prediction. I tried
paddle.device.cuda.empty_cache()
, but it does not work. The GPU memory just keeps increasing as more predictions happen till it reaches 100%. Is there any suggestions? Thanks in advance!Beta Was this translation helpful? Give feedback.
All reactions