Logging of request and response payloads from your Seldon Deployment can be accomplished by adding a logging section to any part of the Seldon deployment graph. An example is shown below:

apiVersion: machinelearning.seldon.io/v1
kind: SeldonDeployment
name: seldon-model
spec:
name: test-deployment
predictors:
- componentSpecs:
- spec:
containers:
- image: seldonio/mock_classifier:1.3
name: classifier
graph:
children: []
endpoint:
type: REST
name: classifier
type: MODEL
logger:
url: http://mylogging-endpoint
mode: all
name: example
replicas: 1


The logging for the top level requets response is provided by:

logger:
url: http://mylogging-endpoint
mode: all


In this example both request and response payloads as specified by the mode attribute are sent as CloudEvents to the url http://mylogging-endpoint.

The specification is:

• url: Any url. Optional. If not provided then it will default to the default knative borker in the namespace of the Seldon Deployment.

• mode: Either request, response or all

## Setting Global Default¶

If you don’t want to set up the custom logger every time, you are able to set it with executor.requestLogger.defaultEndpoint in the Helm Chart Variable as outlined in the helm chart advanced settings section.

This can simply specify a URL to call. In the usual kubernetes fashion, if a service name is provided then it is assumed to be in the current namespace unless there it is followed by .<namespace>, giving the namespace name.

You will still want to make sure the model is deployed with a specification on what requests will be logged, i.e. all, request or response (as outlined above).

### Example Notebook¶

You can try out an example notebook with logging