简体   繁体   English

Triton 推理服务器 - tritonserver:未找到

[英]Triton Inference Server - tritonserver: not found

I try to run NVIDIA's Triton Inference Server.我尝试运行 NVIDIA 的 Triton 推理服务器。 I pulled the pre-built container nvcr.io/nvidia/pytorch:22.06-py3 and then run it with the command我拉了预建的容器 nvcr.io/nvidia/pytorch:22.06-py3 然后用命令运行它

run --gpus=1 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/F/models:/models nvcr.io/nvidia/pytorch:22.06-py3 tritonserver --model-repository=/models

and got the error并得到了错误

/opt/nvidia/nvidia_entrypoint.sh: line 49: exec: tritonserver: not found

I googled and have not found something to catch this.我用谷歌搜索并没有找到可以捕捉到这个的东西。 I tried to change tritonserver to trtserver as recommended but it did not help.我尝试按照建议将 tritonserver 更改为 trtserver ,但没有帮助。 Please give some advice how it can be solved.请给一些建议如何解决它。

Looks like you're trying to run a tritonserver using a pytorch image but according to the triton-server quick start guide , the image should be:看起来您正在尝试使用pytorch图像运行tritonserver ,但根据triton-server quick start guide ,图像应该是:

$ docker run --gpus=1 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/full/path/to/docs/examples/model_repository:/models nvcr.io/nvidia/tritonserver:<xx.yy>-py3 tritonserver --model-repository=/models

Where <xx.yy> is the version of Triton that you want to use其中 <xx.yy> 是您要使用的 Triton 版本

In your case it should be nvcr.io/nvidia/tritonserver:22.06-py3 and the full command:在您的情况下,它应该是nvcr.io/nvidia/tritonserver:22.06-py3和完整的命令:

run --gpus=1 --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/F/models:/models nvcr.io/nvidia/tritonserver:22.06-py3 tritonserver --model-repository=/models

声明:本站的技术帖子网页,遵循CC BY-SA 4.0协议,如果您需要转载,请注明本站网址或者原文地址。任何问题请咨询:yoyou2525@163.com.

 
粤ICP备18138465号  © 2020-2024 STACKOOM.COM