Inference Artifacts
To run your model inside Seldon you must supply an inference artifact that can be downloaded and run on one of MLServer or Triton inference servers. We list artifacts below by alphabetical order below.
Type | Server | Tag | Example |
---|---|---|---|
Alibi-Detect | MLServer |
| |
Alibi-Explain | MLServer |
| |
DALI | Triton |
| TBC |
Huggingface | MLServer |
| |
LightGBM | MLServer |
| |
MLFlow | MLServer |
| |
ONNX | Triton |
| |
OpenVino | Triton |
| TBC |
Custom Python | MLServer |
| |
Custom Python | Triton |
| |
PyTorch | Triton |
| |
SKLearn | MLServer |
| |
Spark Mlib | MLServer |
| TBC |
Tensorflow | Triton |
| |
TensorRT | Triton |
| TBC |
Triton FIL | Triton |
| TBC |
XGBoost | MLServer |
|
Saving Model artifacts
For many machine learning artifacts you can simply save them to a folder and load them into Seldon Core 2. Details are given below as well as a link to creating a custom model settings file if needed.
Type | Notes | Custom Model Settings |
---|---|---|
Alibi-Detect | ||
Alibi-Explain | ||
DALI | Follow the Triton docs to create a config.pbtxt and model folder with artifact. | |
Huggingface | Create an MLServer | |
LightGBM | Save model to file with extension | |
MLFlow | Use the created | |
ONNX | Save you model with name | |
OpenVino | Follow the Triton docs to create your model artifacts. | |
Custom MLServer Python | Create a python file with a class that extends | |
Custom Triton Python | Follow the Triton docs to create your | |
PyTorch | Create a Triton | |
SKLearn | Save model via joblib to a file with extension | |
Spark Mlib | Follow the MLServer docs. | |
Tensorflow | Save model in "Saved Model" format as | |
TensorRT | Follow the Triton docs to create your model artifacts. | |
Triton FIL | Follow the Triton docs to create your model artifacts. | |
XGBoost | Save model to file with extension |
Custom MLServer Model Settings
For MLServer targeted models you can create a model-settings.json
file to help MLServer load your model and place this alongside your artifact. See the MLServer project for details.
Custom Triton Configuration
For Triton inference server models you can create a configuration config.pbtxt file alongside your artifact.
Notes
The tag
field represents the tag you need to add to the requirements
part of the Model spec for your artifact to be loaded on a compatible server. e.g. for an sklearn model:
Last updated