You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Why, after using OneDNN, did I find that the GEMM used in the call stack extracted with perf is arm_gemm from ARM Compute instead of BRGEMM from OneDNN? #2256
When I run TF Serving on an x64 machine, I notice that TensorFlow uses brgemm_matmul_t for inference, while on an ARM architecture machine, it uses arm_gemm. How can I also use brgemm_matmul on ARM, as it provides better performance?