seldon-openvino

:exclamation: This Helm Chart is deprecated!

Version: 0.1

Proxy integration to deploy models optimized for Intel OpenVINO in Seldon Core v1

Usage

To use this chart, you will first need to add the seldonio Helm repo:

helm repo add seldonio https://storage.googleapis.com/seldon-charts
helm repo update

Once that’s done, you should then be able to use the inference graph template as:

helm template $MY_MODEL_NAME seldonio/seldon-openvino --namespace $MODELS_NAMESPACE

Note that you can also deploy the inference graph directly to your cluster using:

helm install $MY_MODEL_NAME seldonio/seldon-openvino --namespace $MODELS_NAMESPACE

Source Code

Values

Key

Type

Default

Description

engine.env.SELDON_LOG_MESSAGES_EXTERNALLY

bool

false

engine.env.SELDON_LOG_REQUESTS

bool

false

engine.env.SELDON_LOG_RESPONSES

bool

false

engine.resources.requests.cpu

string

"0.1"

openvino.image

string

"intelaipg/openvino-model-server:0.3"

openvino.model.env.LOG_LEVEL

string

"DEBUG"

openvino.model.input

string

"data"

openvino.model.name

string

"squeezenet1.1"

openvino.model.output

string

"prob"

openvino.model.path

string

"/opt/ml/squeezenet"

openvino.model.resources

object

{}

openvino.model.src

string

"gs://seldon-models/openvino/squeezenet"

openvino.model_volume

string

"hostPath"

openvino.port

int

8001

predictorLabels.fluentd

string

"true"

predictorLabels.version

string

"v1"

sdepLabels.app

string

"seldon"

tfserving_proxy.image

string

"seldonio/tfserving-proxy:0.2"