# Explaining Autopilot Models

Kernel `Python 3 (Data Science)`

works well with this notebook.

*This notebook was created and tested on an ml.m5.xlarge notebook instance.*

## Table of Contents

## Introduction

Machine learning (ML) models have long been considered black boxes since predictions from these models are hard to interpret. While decision trees can be interpreted by observing the parameters learned by the models, it is generally difficult to get a clear picture.

Model interpretation can be divided into local and global explanations. A local explanation considers a single sample and answers questions like: “why the model predicts that customer A will stop using the product?” or “why the ML system refused John Doe a loan?”. Another interesting question is “what should John Doe change in order to get the loan approved?”. On the contrary, global explanations aim at explaining the model itself and answer questions like “which features are important for prediction?”. It is important to note that local explanations can be used to derive global explanations by averaging many samples. For further reading on interpretable ML, see the excellent book by Christoph Molnar.

In this blog post, we will demonstrate the use of the popular model interpretation framework SHAP for both local and global interpretation.

### SHAP

SHAP is a game theoretic framework inspired by Shapley Values that provides local explanations for any model. SHAP has gained popularity in recent years, probably due to its strong theoretical basis. The SHAP package contains several algorithms that, given a sample and a model, derive the SHAP value for each of the model’s input features. The SHAP value of a feature represents the feature’s contribution to the model’s prediction.

To explain models built by Amazon SageMaker Autopilot we use SHAP’s `KernelExplainer`

which is a black box explainer. `KernelExplainer`

is robust and can explain any model, thus can handle Autopilot’s complex feature processing. `KernelExplainer`

only requires that the model will support an inference functionality which, given a sample, will return the model’s prediction for that sample. The prediction being the predicted value for
regression and the class probability for classification.

It is worth noting that SHAP includes several other explainers such as `TreeExplainer`

and `DeepExplainer`

that are specific for decision forest and neural networks respectively. These are not black box explainers and require knowledge of the model structure and trained params. `TreeExplainer`

and `DeepExplainer`

are limited and currently can not support any feature processing.

## Setup

In this notebook we will start with a model built by SageMaker Autopilot which was already trained on a binary classification task. Please refer to this notebook to see how to create and train an Autopilot model.

```
[ ]:
```

```
import boto3
import pandas as pd
import sagemaker
from sagemaker import AutoML
from datetime import datetime
import numpy as np
region = boto3.Session().region_name
session = sagemaker.Session()
```

Install SHAP

```
[ ]:
```

```
%conda install -c conda-forge shap
```

```
[ ]:
```

```
import shap
from shap import KernelExplainer
from shap import sample
from scipy.special import expit
# Initialize plugin to make plots interactive.
shap.initjs()
```

### Create an inference endpoint

Creating an inference endpoint for the trained Autopilot model. Skip this step if an endpoint with the argument `inference_response_keys`

set as `['predicted_label', 'probability']`

was already created.

```
[ ]:
```

```
automl_job_name = "your-autopilot-job-that-exists"
automl_job = AutoML.attach(automl_job_name, sagemaker_session=session)
# Endpoint name
ep_name = "sagemaker-automl-" + datetime.now().strftime("%Y-%m-%d-%H-%M-%S")
```

```
[ ]:
```

```
# For classification response to work with SHAP we need the probability scores. This can be achieved by providing a list of keys for
# response content. The order of the keys will dictate the content order in the response. This parameter is not needed for regression.
inference_response_keys = ["predicted_label", "probability"]
# Create the inference endpoint
automl_job.deploy(
initial_instance_count=1,
instance_type="ml.m5.2xlarge",
inference_response_keys=inference_response_keys,
endpoint_name=ep_name,
)
```

### Wrap Autopilot’s endpoint with an estimator class.

For ease of use, we wrap the inference endpoint with a custom estimator class. Two inference functions are provided: `predict`

which returns the numeric prediction value to be used for regression and `predict_proba`

which returns the class probability to be used for classification.

```
[ ]:
```

```
from sagemaker.predictor import Predictor
class AutomlEstimator:
def __init__(self, endpoint_name, sagemaker_session):
self.predictor = Predictor(
endpoint_name=endpoint_name,
sagemaker_session=sagemaker_session,
serializer=sagemaker.serializers.CSVSerializer(),
content_type="text/csv",
accept="text/csv",
)
def get_automl_response(self, x):
if x.__class__.__name__ == "ndarray":
payload = ""
for row in x:
payload = payload + ",".join(map(str, row)) + "\n"
else:
payload = x.to_csv(sep=",", header=False, index=False)
return self.predictor.predict(payload).decode("utf-8")
# Prediction function for regression
def predict(self, x):
response = self.get_automl_response(x)
# we get the first column from the response array containing the numeric prediction value (or label in case of classification)
response = np.array([x.split(",")[0] for x in response.split("\n")[:-1]])
return response
# Prediction function for classification
def predict_proba(self, x):
"""Extract and return the probability score from the AutoPilot endpoint response."""
response = self.get_automl_response(x)
# we get the second column from the response array containing the class probability
response = np.array([x.split(",")[1] for x in response.split("\n")[:-1]])
return response.astype(float)
```

Create an instance of `AutomlEstimator`

```
[ ]:
```

```
automl_estimator = AutomlEstimator(endpoint_name=ep_name, sagemaker_session=session)
```

### Data

In this notebook we will use the same dataset as used in the Customer Churn notebook. Please follow the “Customer Churn” notebook to download the dataset if it was not previously downloaded.

### Background data

KernelExplainer requires a sample of the data to be used as background data. KernelExplainer uses this data to simulate a feature being missing by replacing the feature value with a random value from the background. We use shap.sample to sample 50 rows from the dataset to be used as background data. Using more samples as background data will produce more accurate results but runtime will increase. Choosing background data is challenging, see the whitepapers: https://storage.googleapis.com/cloud-ai-whitepapers/AI%20Explainability%20Whitepaper.pdf and https://docs.seldon.io/projects/alibi/en/latest/methods/KernelSHAP.html#Runtime-considerations. Note that the clustering algorithms provided in shap only support numeric data. According to SHAP’s documentation, a vector of zeros could be used as background data to produce reasonable results.

```
[ ]:
```

```
churn_data = pd.read_csv("../churn.txt")
data_without_target = churn_data.drop(columns=["Churn?"])
background_data = sample(data_without_target, 50)
```

### Setup KernelExplainer

Next, we create the `KernelExplainer`

. Note that since it’s a black box explainer, `KernelExplainer`

only requires a handle to the predict (or predict_proba) function and does not require any other information about the model. For classification it is recommended to derive feature importance scores in the log-odds space since additivity is a more natural assumption there thus we use `logit`

. For regression `identity`

should be used.

```
[ ]:
```

```
# Derive link function
problem_type = automl_job.describe_auto_ml_job(job_name=automl_job_name)["ResolvedAttributes"][
"ProblemType"
]
link = "identity" if problem_type == "Regression" else "logit"
# the handle to predict_proba is passed to KernelExplainer since KernelSHAP requires the class probability
explainer = KernelExplainer(automl_estimator.predict_proba, background_data, link=link)
```

By analyzing the background data `KernelExplainer`

provides us with `explainer.expected_value`

which is the model prediction with all features missing. Considering a customer for which we have no data at all (i.e. all features are missing) this should theoretically be the model prediction.

```
[ ]:
```

```
# Since expected_value is given in the log-odds space we convert it back to probability using expit which is the inverse function to logit
print("expected value =", expit(explainer.expected_value))
```

## Local explanation with KernelExplainer

We will use `KernelExplainer`

to explain the prediction of a single sample, the first sample in the dataset.

```
[ ]:
```

```
# Get the first sample
x = data_without_target.iloc[0:1]
# ManagedEndpoint can optionally auto delete the endpoint after calculating the SHAP values. To enable auto delete, use ManagedEndpoint(ep_name, auto_delete=True)
from managed_endpoint import ManagedEndpoint
with ManagedEndpoint(ep_name) as mep:
shap_values = explainer.shap_values(x, nsamples="auto", l1_reg="aic")
```

SHAP package includes many visualization tools. See below a `force_plot`

which provides a good visualization for the SHAP values of a single sample

```
[ ]:
```

```
# Since shap_values are provided in the log-odds space, we convert them back to the probability space by using LogitLink
shap.force_plot(explainer.expected_value, shap_values, x, link=link)
```

From the plot above we learn that the most influential feature is `VMail Message`

which pushes the probability down by about 7%. It is important to note that `VMail Message = 25`

makes the probability 7% lower in comparison to the notion of that feature being missing. SHAP values do not provide the information of how increasing or decreasing `VMail Message`

will affect prediction.

In many cases we are interested only in the most influential features. By setting `l1_reg='num_features(5)'`

, SHAP will provide non-zero scores for only the most influential 5 features.

```
[ ]:
```

```
with ManagedEndpoint(ep_name) as mep:
shap_values = explainer.shap_values(x, nsamples="auto", l1_reg="num_features(5)")
shap.force_plot(explainer.expected_value, shap_values, x, link=link)
```

## KernelExplainer computation cost

KernelExplainer computation cost is dominated by the inference calls. In order to estimate SHAP values for a single sample, KernelExplainer calls the inference function twice: First, with the sample unaugmented. And second, with many randomly augmented instances of the sample. The number of augmented instances in our case is: 50 (#samples in the background data) * 2088 (nsamples = ‘auto’) = 104,400. So, in our case, the cost of running KernelExplainer for a single sample is roughly the cost of 104,400 inference calls.

## Global explanation with KernelExplainer

Next, we will use KernelExplainer to provide insight about the model as a whole. It is done by running KernelExplainer locally on 50 samples and aggregating the results

```
[ ]:
```

```
# Sample 50 random samples
X = sample(data_without_target, 50)
# Calculate SHAP values for these samples, and delete the endpoint
with ManagedEndpoint(ep_name, auto_delete=True) as mep:
shap_values = explainer.shap_values(X, nsamples="auto", l1_reg="aic")
```

`force_plot`

can be used to visualize SHAP values for many samples simultaneously by rotating the plot of each sample by 90 degrees and stacking the plots horizontally. The resulting plot is interactive and can be manually analyzed.

```
[ ]:
```

```
shap.force_plot(explainer.expected_value, shap_values, X, link=link)
```

`summary_plot`

is another visualization tool displaying the mean absolute value of the SHAP values for each feature using a bar plot. Currently, `summary_plot`

does not support link functions so the SHAP values are presented in the log-odds space (and not the probability space).

```
[ ]:
```

```
shap.summary_plot(shap_values, X, plot_type="bar")
```

## Conclusion

In this post, we demonstrated how to use KernelSHAP to explain models created by Amazon SageMaker Autopilot both locally and globally. KernelExplainer is a robust black box explainer which requires only that the model will support an inference functionality which, given a sample, returns the model’s prediction for that sample. This inference functionality was provided by wrapping Autopilot’s inference endpoint with an estimator container.

For more about Amazon SageMaker Autopilot, please see Amazon SageMaker Autopilot.