当前位置:   article > 正文

Triton 模型推理使用案例、tritonclient https,grpc访问;导出onnx指定输入维度可变dynamic_axes_triton client

triton client

参考:https://www.cnblogs.com/infgrad/p/16283130.html
http://t.zoukankan.com/zzk0-p-15543824.html
https://www.bilibili.com/video/BV1ET411G7zV

1、下载安装Triton docker版

参考:https://github.com/triton-inference-server/server/blob/main/docs/quickstart.md

docker pull nvcr.io/nvidia/tritonserver:<xx.yy>-py3
  • 1

**我测试安装的是nvcr.io/nvidia/tritonserver:20.09-py3

2、运行案例

# Step 1: Create the example model repository 
git clone  https://github.com/triton-inference-server/server.git

cd server/docs/examples

./fetch_models.sh

# Step 2: Launch triton from the NGC Triton container
docker run --gpus=1 --rm --net=host -v /full/path/to/docs/examples/model_repository:/models nvcr.io/nvidia/tritonserver:20.09-py3 tritonserver --model-repository=/models

  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10

声明:本文内容由网友自发贡献,不代表【wpsshop博客】立场,版权归原作者所有,本站不承担相应法律责任。如您发现有侵权的内容,请联系我们。转载请注明出处:https://www.wpsshop.cn/w/秋刀鱼在做梦/article/detail/885522
推荐阅读
相关标签
  

闽ICP备14008679号