MLServer has been built with Multi-Model Serving (MMS) in mind. This means that, within a single instance of MLServer, you can serve multiple models under different paths. This also includes multiple versions of the same model.
This notebook shows an example of how you can leverage MMS with MLServer.
Training
We will first start by training 2 different models:
# Original source code and more details can be found in:# https://scikit-learn.org/stable/auto_examples/classification/plot_digits_classification.html# Import datasets, classifiers and performance metricsfrom sklearn import datasets, svm, metricsfrom sklearn.model_selection import train_test_split# The digits datasetdigits = datasets.load_digits()# To apply a classifier on this data, we need to flatten the image, to# turn the data in a (samples, feature) matrix:n_samples =len(digits.images)data = digits.images.reshape((n_samples, -1))# Create a classifier: a support vector classifierclassifier = svm.SVC(gamma=0.001)# Split data into train and test subsetsX_train, X_test_digits, y_train, y_test_digits =train_test_split( data, digits.target, test_size=0.5, shuffle=False)# We learn the digits on the first half of the digitsclassifier.fit(X_train, y_train)
The next step will be serving both our models within the same MLServer instance. For that, we will just need to create a model-settings.json file local to each of our models and a server-wide settings.json. That is,
settings.json: holds the configuration of our server (e.g. ports, log level, etc.).
models/mnist-svm/model-settings.json: holds the configuration specific to our mnist-svm model (e.g. input type, runtime to use, etc.).
models/mushroom-xgboost/model-settings.json: holds the configuration specific to our mushroom-xgboost model (e.g. input type, runtime to use, etc.).
Now that we have our config in-place, we can start the server by running mlserver start .. This needs to either be ran from the same directory where our config files are or pointing to the folder where they are.
mlserverstart.
Since this command will start the server and block the terminal, waiting for requests, this will need to be ran in the background on a separate terminal.
Testing
By this point, we should have both our models getting served by MLServer. To make sure that everything is working as expected, let's send a request from each test set.
For that, we can use the Python types that the mlserver package provides out of box, or we can build our request manually.