TensorFlow Serving

If you have a trained Tensorflow model you can deploy this directly via REST or gRPC servers.

MNIST Example

REST MNIST Example

For REST you need to specify parameters for:

  • signature_name

  • model_name

apiVersion: machinelearning.seldon.io/v1alpha2
kind: SeldonDeployment
metadata:
  name: tfserving
spec:
  name: mnist
  predictors:
  - graph:
      children: []
      implementation: TENSORFLOW_SERVER
      modelUri: gs://seldon-models/tfserving/mnist-model
      name: mnist-model
      parameters:
        - name: signature_name
          type: STRING
          value: predict_images
        - name: model_name
          type: STRING
          value: mnist-model
    name: default
    replicas: 1

gRPC MNIST Example

For gRPC you need to specify the following parameters:

  • signature_name

  • model_name

  • model_input

  • model_output

Try out a worked notebook

Multi-Model Serving

You can utilize Tensorflow Serving's functionality to load multiple models from one model repository as shown in this example notebook. You should follow the configuration details as disucussed in the Tensorflow Serving documentation on advanced configuration.

Last updated

Was this helpful?