Inference Runtimes
Last updated
Was this helpful?
Last updated
Was this helpful?
Inference runtimes allow you to define how your model should be used within MLServer. You can think of them as the backend glue between MLServer and your machine learning framework of choice.
Out of the box, MLServer comes with a set of pre-packaged runtimes which let you interact with a subset of common ML frameworks. This allows you to start serving models saved in these frameworks straight away. To avoid bringing in dependencies for frameworks that you don't need to use, these runtimes are implemented as independent (and optional) Python packages. This mechanism also allows you to rollout your own custom runtimes very easily.
To pick which runtime you want to use for your model, you just need to make sure that the right package is installed, and then point to the correct runtime class in your model-settings.json
file.
Scikit-Learn
mlserver-sklearn
mlserver_sklearn.SKLearnModel
XGBoost
mlserver-xgboost
mlserver_xgboost.XGBoostModel
Spark MLlib
mlserver-mllib
mlserver_mllib.MLlibModel
LightGBM
mlserver-lightgbm
mlserver_lightgbm.LightGBMModel
CatBoost
mlserver-catboost
mlserver_catboost.CatboostModel
MLflow
mlserver-mlflow
mlserver_mlflow.MLflowRuntime
Alibi-Detect
mlserver-alibi-detect
mlserver_alibi_detect.AlibiDetectRuntime