NVIDIA TensorRT MNIST

digit

This example shows how you can deploy a TensorRT model with NVIDIA Triton Server. In this case we use a prebuilt TensorRT model for NVIDIA v100 GPUs.

Note this example requires some advanced setup and is directed for those with tensorRT experience.

Prerequisites

  • Install requirements in requirements.txt

  • An authorized kubernetes cluster with V100 GPUs installed and configured.

  • Install Seldon Core and install Ambassador and port-forward to Ambassador on localhost:8003

This example uses the KFServing protocol supported by Triton Infernence Server which Seldon also supports.

Check metadata of model

Test prediction on random digit.

png

Last updated

Was this helpful?