Servers
Learn how to configure and manage inference servers in Seldon Core 2, including MLServer and Triton server farms, model scheduling, and capability management.
apiVersion: mlops.seldon.io/v1alpha1
kind: Model
metadata:
name: iris
spec:
storageUri: "gs://seldon-models/scv2/samples/mlserver_1.5.0/iris-sklearn"
requirements:
- sklearn
memory: 100Ki- name: SELDON_SERVER_CAPABILITIES value: "mlserver,alibi-detect,alibi-explain,huggingface,lightgbm,mlflow,python,sklearn,spark-mlib,xgboost"- name: SELDON_SERVER_CAPABILITIES value: "triton,dali,fil,onnx,openvino,python,pytorch,tensorflow,tensorrt"
Custom Capabilities
Autoscaling of Servers
Last updated
Was this helpful?

