TensorFlow Serving
If you have a trained Tensorflow model you can deploy this directly via REST or gRPC servers.
MNIST Example
REST MNIST Example
For REST you need to specify parameters for:
signature_name
model_name
apiVersion: machinelearning.seldon.io/v1alpha2
kind: SeldonDeployment
metadata:
name: tfserving
spec:
name: mnist
predictors:
- graph:
children: []
implementation: TENSORFLOW_SERVER
modelUri: gs://seldon-models/tfserving/mnist-model
name: mnist-model
parameters:
- name: signature_name
type: STRING
value: predict_images
- name: model_name
type: STRING
value: mnist-model
name: default
replicas: 1
gRPC MNIST Example
For gRPC you need to specify the following parameters:
signature_name
model_name
model_input
model_output
Try out a worked notebook
Multi-Model Serving
You can utilize Tensorflow Serving's functionality to load multiple models from one model repository as shown in this example notebook. You should follow the configuration details as disucussed in the Tensorflow Serving documentation on advanced configuration.
Last updated
Was this helpful?