Replies: 1 comment 2 replies
-
Hi @Saigut, it is required to launch Triton with the script |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
The
./scripts/launch_triton_server.py
in tensorrtllm_backend repo has a parameter--world_size
, buttritonserver
don't.So when I try to launch a "tp_size=4 pp_size=1" tensorrtllm model using
tritonserver
, it will report error like mpiSize != tp_size*pp_size, due to mpiNUm is always 1.Is there any method for
tritonserver
to specifyworld_size
for tensorrtllm model?I think the most proper way maybe to set
world_size
in model repo of tensorrtllm model, like inconfig.pbtxt
, but I can't find such a option.Beta Was this translation helpful? Give feedback.
All reactions