Skip to content

Why, after using OneDNN, did I find that the GEMM used in the call stack extracted with perf is arm_gemm from ARM Compute instead of BRGEMM from OneDNN? #2256

Open
@nanzh-19

Description

@nanzh-19

When I run TF Serving on an x64 machine, I notice that TensorFlow uses brgemm_matmul_t for inference, while on an ARM architecture machine, it uses arm_gemm. How can I also use brgemm_matmul on ARM, as it provides better performance?

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions