How to change the service's batch-size #369
TOMATODA
announced in
Announcements
Replies: 1 comment 2 replies
-
You can you |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I already deployed the model service in the container successfully using the default batch-size , I want to increase the batch-size,what should I do to accomplish it without launching the container once again?
deploy command as follows:
/app/.venv/bin/python /app/.venv/bin/infinity_emb --model-name-or-path /opt/llm/input/pretrain/BAAI--bge-reranker-v2-m3 --port 8002 --device cuda --dtype float16
Beta Was this translation helpful? Give feedback.
All reactions