You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I use the OpenVino Execution Provider (EP) in ONNX Runtime for CPU-based inference, the memory is not de-allocated between inference sessions, causing a memory leak for inputs of various dimensions (model uses dynamic-sized inputs).
After reading #11667, I've tried to build OpenVino checking out the commit that introduced usstq@3857559, but this did not resolve the memory leak.
Steps to reproduce
Build ONNXRuntime 1.11 or 1.12 with --use_openvino AUTO:CPU,GPU
Load the CRAFT model using ONNXRuntime with OpenVino EP
Perform various inference requests (via Nvidia Triton Inference Server)
Issue submission checklist
I report the issue, it's not a question
I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution
There is reproducer code and related data files: images, videos, models, etc.
The text was updated successfully, but these errors were encountered:
System information (version)
Detailed description
When I use the OpenVino Execution Provider (EP) in ONNX Runtime for CPU-based inference, the memory is not de-allocated between inference sessions, causing a memory leak for inputs of various dimensions (model uses dynamic-sized inputs).
After reading #11667, I've tried to build OpenVino checking out the commit that introduced usstq@3857559, but this did not resolve the memory leak.
Steps to reproduce
--use_openvino AUTO:CPU,GPU
Issue submission checklist
The text was updated successfully, but these errors were encountered: