LogoLogo
  • MLServer
  • Getting Started
  • User Guide
    • Content Types (and Codecs)
    • OpenAPI Support
    • Parallel Inference
    • Adaptive Batching
    • Custom Inference Runtimes
    • Metrics
    • Deployment
      • Seldon Core
      • KServe
    • Streaming
  • Inference Runtimes
    • SKLearn
    • XGBoost
    • MLFlow
    • Spark MlLib
    • LightGBM
    • Catboost
    • Alibi-Detect
    • Alibi-Explain
    • HuggingFace
    • Custom
  • Reference
    • MLServer Settings
    • Model Settings
    • MLServer CLI
    • Python API
      • MLModel
      • Types
      • Codecs
      • Metrics
  • Examples
    • Serving Scikit-Learn models
    • Serving XGBoost models
    • Serving LightGBM models
    • Serving MLflow models
    • Serving a custom model
    • Serving Alibi-Detect models
    • Serving HuggingFace Transformer Models
    • Multi-Model Serving
    • Model Repository API
    • Content Type Decoding
    • Custom Conda environments in MLServer
    • Serving a custom model with JSON serialization
    • Serving models through Kafka
    • Streaming
    • Deploying a Custom Tensorflow Model with MLServer and Seldon Core
  • Changelog
Powered by GitBook
On this page
  • Inference Runtimes
  • MLServer Features
  • Tutorials

Was this helpful?

Edit on GitHub
Export as PDF

Examples

PreviousMetricsNextServing Scikit-Learn models

Last updated 7 months ago

Was this helpful?

To see MLServer in action you can check out the examples below. These are end-to-end notebooks, showing how to serve models with MLServer.

Inference Runtimes

If you are interested in how MLServer interacts with particular model frameworks, you can check the following examples. These focus on showcasing the different that ship with MLServer out of the box. Note that, for advanced use cases, you can also write your own custom inference runtime (see the ).

MLServer Features

To see some of the advanced features included in MLServer (e.g. multi-model serving), check out the examples below.

Tutorials

Tutorials are designed to be beginner-friendly and walk through accomplishing a series of tasks using MLServer (and other tools).

inference runtimes
example below on custom models
Serving Scikit-Learn models
Serving XGBoost models
Serving LightGBM models
Serving CatBoost models
Serving MLflow models
Serving custom models
Serving Alibi Detect models
Serving HuggingFace models
Multi-Model Serving with multiple frameworks
Loading / unloading models from a model repository
Content-Type Decoding
Custom Conda environment
Serving custom models requiring JSON inputs or outputs
Serving models through Kafka
Streaming inference
Deploying a Custom Tensorflow Model with MLServer and Seldon Core