I try to run NVIDIA’s Triton Inference Server. I pulled the pre-built container nvcr.io/nvidia/pytorch:22.06-py3 and then run it with the command
run --gpus=1 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/F/models:/models nvcr.io/nvidia/pytorch:22.06-py3 tritonserver --model-repository=/models
and got the error
/opt/nvidia/nvidia_entrypoint.sh: line 49: exec: tritonserver: not found
I googled and have not found something to catch this. I tried to change tritonserver to trtserver as recommended but it did not help. Please give some advice how it can be solved.