Batch Processing with Kubeflow Pipelines

In this notebook we will dive into how you can run batch processing with Kubeflow Pipelines and Seldon Core.

Dependencies:

  • Seldon core installed as per the docs with Istio Ingress

  • Kubeflow Pipelines installed (installation instructions in this notebook)

kubeflow-pipeline

Kubeflow Pipelines Setup

Setup the pipeline in your current cluster:

We also install the Python Library so we can create our pipeline:

Add Batch Data

In order to run our batch job we will need to create some batch data that can be used to process.

This batch dataset will be pushed to a minio instance so it can be downloaded from Minio (which we need to install first)

Install Minio

Forward the Minio port so you can access it

You can do this by runnning the following command in your terminal:

Create some input for our model

We will create a file that will contain the inputs that will be sent to our model

Check the contents of the file

Upload the file to our minio

Create Kubeflow Pipeline

We are now able to create a kubeflow pipeline that will allow us to enter the batch parameters through the UI.

We will also be able to add extra steps that will download the data from a Minio client.

We use the pipeline syntax to create the kubeflow pipeline, as outlined below:

Trigger the creation

We will run the python file which triggers the creation of the pipeline that we can the upload on the UI:

Check the pipeline has been created:

Open the Kubeflow Pipelines UI

We can now open the UI by port forwarding the UI with the following command:

And we can open it locally in our browser via http://localhost:8000

Now we can follow the standard steps to create and deploy the kubeflow pipline

seldon-kubeflow-batch

Last updated

Was this helpful?