Using Openshift Source-to-Image to facilitate Machine Learning Deployment¶
Seldon aims to help organisations put their data science projects into production so they can decrease the time to get return on investment. By helping data scientists take their data science models and place them into production, scale them, get analytics and modify them Seldon allows data scientists to bridge the gap from development to production and use current dev-ops best practices in machine learning. Our core products run on top of Kubernetes and can be deployed on-cloud on on-premise. Integrating with enterprise ready Kubernetes distributions such as Openshift allows us to provide a solid foundation in which to supply our products for use in demanding verticals such as the FinTech sector.
Seldon-Core is an open source project that provides scalable machine learning deployment running on Kubernetes. One of Seldon-Core’s goals is to allow data scientists to continue to construct their training and inference components using any of the many available machine learning toolkits, be that python based (e.g., TensorFlow, sklearn), R or Java (e.g., Spark, H2O) amongst many popular options. Seldon-Core will then allow them easily to package and run their runtime prediction modules on Kubernetes. To achieve this goal we need to make it easy for data scientists to take their source code and package it as a Docker-formatted container in the correct form such that it can be managed as part of a runtime microservice graph on Kubernetes by Seldon-Core. For this we utilize Openshift’s Source-to-Image open source library to allow any code to be packaged in the correct format with minimal requirements from the data scientist.
Seldon-core provides scalable machine learning deployments running on Kubernetes. To deploy their models data scientists follow the steps as shown below:
Package their runtime model as a Docker-formatted image
Describe their runtime graph as a Kubernetes resource
Deploy to Kubernetes using standard tools such as kubectl, Helm, ksonnet.
Once running their deployment can be updated as new image releases are created for the runtime model as well as updates to the runtime graph.
The components of the runtime graph can be of various types. The most typical is a model which will provide predictions given some input features. Typically, the data scientist will have trained a model and saved the model parameters for use by a runtime component that will be provide new predictions at runtime. However, Seldon-Core allows a range of components to be created that can be joined together as building blocks to create more complex runtime graphs as show below:
The types of component you can create can include:
Models - e.g., TensorFlow, sklearn models
Routers - e.g., A-B Tests, Multi-Armed Bandits
Combiners - e.g., Model ensemblers
Transformers - e.g., Feature normalization, Outlier detection, concept drift
As the above diagram shows these need to be fitted into the microservice API of seldon-core either as REST or gRPC services.
To integrate a component into seldon-core the data scientist needs to accomplish two things:
Create a Docker-formatted image from your source code
Wrap your component as a service that exposes REST or gRPC endpoints that follow the seldon-core miroserice API.
To accomplish this we use Openshift’s source-to-image (s2i) open source tool. S2i allows data scientists to wrap their code using a single command line call that can easily be embedded into a continuous integration pipeline. Seldon provides s2i builder images that contain middleware code to wrap the data scientist’s component within a REST or gRPC server that respects the seldon-core microservice API. All that is needed is for the data scientist to follow a few conventions when creating their component in various languages as will be illustrated below. The growing set of source-to-image builder images can be found here.
There are many popular machine learning libraries in python including Tensorflow, keras, sklearn, pyTorch and Statsmodels amongst many others. To use the Seldon-Core s2i builder image to package a python model the data scientist simply needs to provide:
A python file with a class that runs your model
optional requirements.txt or setup.py
.s2i/environment - model definitions used by the s2i builder to correctly wrap your model
The data scientist’s source code should contain a python file which defines a class of the same name as the file. For example:
class MyModel(object): """ Model template. You can load your model parameters in __init__ from a location accessible at runtime """ def __init__(self): """ Add any initialization parameters. These will be passed at runtime from the graph definition parameters defined in your seldondeployment Kubernetes resource manifest. """ print("Initializing") def predict(self,X,features_names): """ Return a prediction. Parameters ---------- X : array-like feature_names : array of feature names (optional) """ print("Predict called - will run identity function") return X
The file is called MyModel.py and it defines a class MyModel
The class contains a predict method that takes an array (numpy) X and feature_names and returns an array of predictions.
Any required initialization can be put inside the class init method.
An optional requirements.txt can detail any software dependencies the code requires.
To allow the s2i builder image to correctly package the component the data scientist needs to provide a few environment variables either in an .s2i/environment file in the source code folder or on the command line. An example is:
Finally we Use
s2i build to create the Docker-formatted image from source code. Examples for python3 code are:
s2i build <git-repo> seldonio/seldon-core-s2i-python3 <my-image-name> s2i build <src-folder> seldonio/seldon-core-s2i-python3 <my-image-name>
By utilizing Openshift’s source-to-image tool data scientists can easily build Docker-formatted images for their runtime components to be deployed at scale using seldon-core. This allows data science teams to use the best machine learning tool for the task and deploy the resulting model in a consistent manner. The seldon-core project is working on providing full Openshift integration in the near future so that Enterprise customers can easily utilize machine learning models within their organisation.
Seldon will be joining Openshift Commons and will be present at Kubecon Europe 2018 and the OpenShift Kubecon Europe event on Tues 1st May. Feel free to contact us to discuss Seldon-Core and Openshift and how they can work together to help data scientists put machine learning into production.