How to change the service's batch-size #369
TOMATODA
announced in
Announcements
Replies: 1 comment 2 replies
-
|
You can you |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I already deployed the model service in the container successfully using the default batch-size , I want to increase the batch-size,what should I do to accomplish it without launching the container once again?
deploy command as follows:
/app/.venv/bin/python /app/.venv/bin/infinity_emb --model-name-or-path /opt/llm/input/pretrain/BAAI--bge-reranker-v2-m3 --port 8002 --device cuda --dtype float16
Beta Was this translation helpful? Give feedback.
All reactions