githubEdit

TensorFlow Serving

If you have a trained Tensorflow model you can deploy this directly via REST or gRPC servers.

MNIST Example

REST MNIST Example

For REST you need to specify parameters for:

  • signature_name

  • model_name

apiVersion: machinelearning.seldon.io/v1alpha2
kind: SeldonDeployment
metadata:
  name: tfserving
spec:
  name: mnist
  predictors:
  - graph:
      children: []
      implementation: TENSORFLOW_SERVER
      modelUri: gs://seldon-models/tfserving/mnist-model
      name: mnist-model
      parameters:
        - name: signature_name
          type: STRING
          value: predict_images
        - name: model_name
          type: STRING
          value: mnist-model
    name: default
    replicas: 1

gRPC MNIST Example

For gRPC you need to specify the following parameters:

  • signature_name

  • model_name

  • model_input

  • model_output

Try out a worked notebookarrow-up-right

Multi-Model Serving

You can utilize Tensorflow Serving's functionality to load multiple models from one model repository as shown in this example notebookarrow-up-right. You should follow the configuration details as disucussed in the Tensorflow Serving documentation on advanced configurationarrow-up-right.

Last updated

Was this helpful?