Skip to content

[Performance] Can oneDNN EP accelerate the inference time of onnxruntime on x86 machines? #14749

Open
@sanbuphy

Description

Describe the issue

I would like to ask the difference between the default CPU EP and oneDNN EP, whether the oneDNN EP can accelerate inference time at the operator level?

I tried openvino EP, but it doesn't work well for dynamic shapes input such as nlp tasks, not as good as the default CPU EP.

To reproduce

I want to speed up the effect of onnxruntime on x86 machine

Urgency

No response

Platform

Linux

OS Version

20.04

ONNX Runtime Installation

Released Package

ONNX Runtime Version or Commit ID

lasted

ONNX Runtime API

Python

Architecture

X64

Execution Provider

Default CPU

Execution Provider Library Version

No response

Model File

No response

Is this a quantized model?

No

Thank you very much.

Metadata

Assignees

No one assigned

    Labels

    ep:OpenVINOissues related to OpenVINO execution providerep:oneDNNquestions/issues related to DNNL EP

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions