NVIDIA Triton

Run TensorRT, ONNX, PyTorch and Tenorflow models on GPUs with NVIDIA Triton

NVIDIA Triton Server

We will:

  • Deploy an image classification model on NVIDIA Triton with GPUs

Deploy Model

For this example choose cifar10 as the name and use the KFServing protocol option.

wizard1

For the model to run we have created several image classification models from the CIFAR10 dataset.

  • Tensorflow Resnet32 model: gs://seldon-models/triton/tf_cifar10
  • ONNX model: gs://seldon-models/triton/onnx_cifar10
  • PyTorch Torchscript model: gs://seldon-models/triton/pytorch_cifar10

Choose one of these and select Triton as the server. Customize the model name to that of the name of the model saved in the bucket for Triton to load.

wizard2

Next, on the resources screen add 1 GPU request/limit assuming you have these available on your cluster and ensure your have provided enough memory for the model. To determine these settings we recommend you use the NVIDIA model analyzer.

wizard3

When ready you can test with images. The payload will depend on the model from above you launched.


Last modified March 3, 2021