Open
Description
Describe the issue
I would like to ask the difference between the default CPU EP and oneDNN EP, whether the oneDNN EP can accelerate inference time at the operator level?
I tried openvino EP, but it doesn't work well for dynamic shapes input such as nlp tasks, not as good as the default CPU EP.
To reproduce
I want to speed up the effect of onnxruntime on x86 machine
Urgency
No response
Platform
Linux
OS Version
20.04
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
lasted
ONNX Runtime API
Python
Architecture
X64
Execution Provider
Default CPU
Execution Provider Library Version
No response
Model File
No response
Is this a quantized model?
No
Thank you very much.