Seldon Core Workflow¶
There are 3 steps to using seldon-core.
- Install seldon-core onto a Kubernetes cluster
- Wrap your components (usually runtime model servers) as Docker containers that respect the internal Seldon microservice API.
- Define your runtime service graph as a SeldonDeployment resource and deploy your model and serve predictions
Wrap Your Model¶
The components you want to run in production need to be wrapped as Docker containers that respect the Seldon microservice API. You can create models that serve predictions, routers that decide on where requests go, such as A-B Tests, Combiners that combine responses and transformers that provide generic components that can transform requests and/or responses.
To allow users to easily wrap machine learning components built using different languages and toolkits we provide wrappers that allow you easily to build a docker container from your code that can be run inside seldon-core. Our current recommended tool is RedHat’s Source-to-Image. More detail can be found in Wrapping your models docs.
Define Runtime Service Graph¶
To run your machine learning graph on Kubernetes you need to define how the components you created in the last step fit together to represent a service graph. This is defined inside a
SeldonDeployment Kubernetes Custom resource. A guide to constructing this inference graph is provided.