Seldon Core Workflow

There are 3 steps to using seldon-core.

  1. Install seldon-core onto a Kubernetes cluster
  2. Wrap your components (usually runtime model servers) as Docker containers that respect the internal Seldon microservice API.
  3. Define your runtime service graph as a SeldonDeployment resource and deploy your model and serve predictions

Install Seldon Core

To install seldon-core follow the installation guide.

Wrap Your Model

The components you want to run in production need to be wrapped as Docker containers that respect the Seldon microservice API. You can create models that serve predictions, routers that decide on where requests go, such as A-B Tests, Combiners that combine responses and transformers that provide generic components that can transform requests and/or responses.

To allow users to easily wrap machine learning components built using different languages and toolkits we provide wrappers that allow you easily to build a docker container from your code that can be run inside seldon-core. Our current recommended tool is RedHat’s Source-to-Image. More detail can be found in Wrapping your models docs.

Define Runtime Service Graph

To run your machine learning graph on Kubernetes you need to define how the components you created in the last step fit together to represent a service graph. This is defined inside a SeldonDeployment Kubernetes Custom resource. A guide to constructing this inference graph is provided.


Deploy and Serve Predictions

You can use kubectl to deploy your ML service like any other Kubernetes resource. This is discussed here. Once deployed ypu can get predictions by calling the exposed API.

Next Steps

Run a notebook using Helm that illustrates using our Helm charts for launching various types of inference graphs.