Open
Description
Description
Using the same model as in #102, the Triton Inference Server has a memory leak, as observed by docker stats
, after adding:
execution_accelerators {
cpu_execution_accelerator : [ {
name : "openvino"
} ]
}
to model config.
Without the openvino
EP usage, there is no memory leak
Triton Information
What version of Triton are you using?
openvino==2022.1.0
with triton-onnxbackend==22.06
and onnxruntime==1.11.1
.
Are you using the Triton container or did you build it yourself?
Custom container build.
To Reproduce
See #102 for model.
Expected behavior
A clear and concise description of what you expected to happen.
Provision of model configuration flags (like in #102) that will customize the memory handling of OpenVino EP.
Metadata
Metadata
Assignees
Labels
No labels