how to use benchmark_app test qwen2.5-7B-instruct model inference performance? #29907
Answered
by
ilya-lavrenov
Light-Travlling
asked this question in
Q&A
-
Beta Was this translation helpful? Give feedback.
Answered by
ilya-lavrenov
Apr 3, 2025
Replies: 2 comments 2 replies
-
Please, use a dedicated benchmark for LLMs from GenAI repo |
Beta Was this translation helpful? Give feedback.
1 reply
-
For multi request, you need to use continuous matching to maximize tput https://github.com/openvinotoolkit/openvino.genai/tree/master/tools/continuous_batching/benchmark |
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
Light-Travlling
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
For multi request, you need to use continuous matching to maximize tput https://github.com/openvinotoolkit/openvino.genai/tree/master/tools/continuous_batching/benchmark