Orchestrate Jobs to Train and Evaluate Models with Amazon SageMaker Pipelines
Amazon SageMaker Pipelines offers machine learning (ML) application developers and operations engineers the ability to orchestrate SageMaker jobs and author reproducible ML pipelines. It also enables them to deploy custom-built models for inference in real-time with low latency, run offline inferences with Batch Transform, and track lineage of artifacts. They can institute sound operational practices in deploying and monitoring production workflows, deploying model artifacts, and tracking artifact lineage through a simple interface, adhering to safety and best practice paradigms for ML application development.
The SageMaker Pipelines service supports a SageMaker Pipeline domain specific language (DSL), which is a declarative JSON specification. This DSL defines a directed acyclic graph (DAG) of pipeline parameters and SageMaker job steps. The SageMaker Python Software Developer Kit (SDK) streamlines the generation of the pipeline DSL using constructs that engineers and scientists are already familiar with.
Runtime
This notebook takes approximately an hour to run.
Contents
SageMaker Pipelines
SageMaker Pipelines supports the following activities, which are demonstrated in this notebook:
Pipelines - A DAG of steps and conditions to orchestrate SageMaker jobs and resource creation.
Processing job steps - A simplified, managed experience on SageMaker to run data processing workloads, such as feature engineering, data validation, model evaluation, and model interpretation.
Training job steps - An iterative process that teaches a model to make predictions by presenting examples from a training dataset.
Conditional execution steps - A step that provides conditional execution of branches in a pipeline.
Register model steps - A step that creates a model package resource in the Model Registry that can be used to create deployable models in Amazon SageMaker.
Create model steps - A step that creates a model for use in transform steps or later publication as an endpoint.
Transform job steps - A batch transform to preprocess datasets to remove noise or bias that interferes with training or inference from a dataset, get inferences from large datasets, and run inference when a persistent endpoint is not needed.
Fail steps - A step that stops a pipeline execution and marks the pipeline execution as failed.
Parametrized Pipeline executions - Enables variation in pipeline executions according to specified parameters.
Notebook Overview
This notebook shows how to:
Define a set of Pipeline parameters that can be used to parametrize a SageMaker Pipeline.
Define a Processing step that performs cleaning, feature engineering, and splitting the input data into train and test data sets.
Define a Training step that trains a model on the preprocessed train data set.
Define a Processing step that evaluates the trained model’s performance on the test dataset.
Define a Create Model step that creates a model from the model artifacts used in training.
Define a Transform step that performs batch transformation based on the model that was created.
Define a Register Model step that creates a model package from the estimator and model artifacts used to train the model.
Define a Conditional step that measures a condition based on output from prior steps and conditionally executes other steps.
Define a Fail step with a customized error message indicating the cause of the execution failure.
Define and create a Pipeline definition in a DAG, with the defined parameters and steps.
Start a Pipeline execution and wait for execution to complete.
Download the model evaluation report from the S3 bucket for examination.
Start a second Pipeline execution.
A SageMaker Pipeline
The pipeline that you create follows a typical machine learning (ML) application pattern of preprocessing, training, evaluation, model creation, batch transformation, and model registration:
Dataset
The dataset you use is the UCI Machine Learning Abalone Dataset [1]. The aim for this task is to determine the age of an abalone snail from its physical measurements. At the core, this is a regression problem.
The dataset contains several features: length (the longest shell measurement), diameter (the diameter perpendicular to length), height (the height with meat in the shell), whole_weight (the weight of whole abalone), shucked_weight (the weight of meat), viscera_weight (the gut weight after bleeding), shell_weight (the weight after being dried), sex (‘M’, ‘F’, ‘I’ where ‘I’ is Infant), and rings (integer).
The number of rings turns out to be a good approximation for age (age is rings + 1.5). However, to obtain this number requires cutting the shell through the cone, staining the section, and counting the number of rings through a microscope, which is a time-consuming task. However, the other physical measurements are easier to determine. You use the dataset to build a predictive model of the variable rings through these other physical measurements.
Before you upload the data to an S3 bucket, upgrade the Amazon SageMaker Python SDK to the latest version and gather some constants you can use later in this notebook.
[1] Dua, D. and Graff, C. (2019). UCI Machine Learning Repository. Irvine, CA: University of California, School of Information and Computer Science.
[2]:
!pip install --upgrade sagemaker
import boto3
import sagemaker
sagemaker_session = sagemaker.session.Session()
region = sagemaker_session.boto_region_name
role = sagemaker.get_execution_role()
default_bucket = sagemaker_session.default_bucket()
model_package_group_name = f"AbaloneModelPackageGroupName"
/opt/conda/lib/python3.7/site-packages/secretstorage/dhcrypto.py:16: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead
from cryptography.utils import int_from_bytes
/opt/conda/lib/python3.7/site-packages/secretstorage/util.py:25: CryptographyDeprecationWarning: int_from_bytes is deprecated, use int.from_bytes instead
from cryptography.utils import int_from_bytes
Requirement already satisfied: sagemaker in /opt/conda/lib/python3.7/site-packages (2.69.1.dev0)
Collecting sagemaker
Downloading sagemaker-2.86.2.tar.gz (521 kB)
|████████████████████████████████| 521 kB 8.6 MB/s
Collecting attrs==20.3.0
Downloading attrs-20.3.0-py2.py3-none-any.whl (49 kB)
|████████████████████████████████| 49 kB 8.4 MB/s
Collecting boto3>=1.20.21
Downloading boto3-1.21.42-py3-none-any.whl (132 kB)
|████████████████████████████████| 132 kB 71.1 MB/s
Requirement already satisfied: google-pasta in /opt/conda/lib/python3.7/site-packages (from sagemaker) (0.2.0)
Requirement already satisfied: numpy>=1.9.0 in /opt/conda/lib/python3.7/site-packages (from sagemaker) (1.21.1)
Requirement already satisfied: protobuf>=3.1 in /opt/conda/lib/python3.7/site-packages (from sagemaker) (3.17.3)
Requirement already satisfied: protobuf3-to-dict>=0.1.5 in /opt/conda/lib/python3.7/site-packages (from sagemaker) (0.1.5)
Requirement already satisfied: smdebug_rulesconfig==1.0.1 in /opt/conda/lib/python3.7/site-packages (from sagemaker) (1.0.1)
Requirement already satisfied: importlib-metadata>=1.4.0 in /opt/conda/lib/python3.7/site-packages (from sagemaker) (1.5.0)
Requirement already satisfied: packaging>=20.0 in /opt/conda/lib/python3.7/site-packages (from sagemaker) (20.1)
Requirement already satisfied: pandas in /opt/conda/lib/python3.7/site-packages (from sagemaker) (1.0.1)
Requirement already satisfied: pathos in /opt/conda/lib/python3.7/site-packages (from sagemaker) (0.2.8)
Requirement already satisfied: jmespath<2.0.0,>=0.7.1 in /opt/conda/lib/python3.7/site-packages (from boto3>=1.20.21->sagemaker) (0.10.0)
Collecting botocore<1.25.0,>=1.24.42
Downloading botocore-1.24.42-py3-none-any.whl (8.7 MB)
|████████████████████████████████| 8.7 MB 51.4 MB/s
Requirement already satisfied: s3transfer<0.6.0,>=0.5.0 in /opt/conda/lib/python3.7/site-packages (from boto3>=1.20.21->sagemaker) (0.5.0)
Requirement already satisfied: urllib3<1.27,>=1.25.4 in /opt/conda/lib/python3.7/site-packages (from botocore<1.25.0,>=1.24.42->boto3>=1.20.21->sagemaker) (1.26.6)
Requirement already satisfied: python-dateutil<3.0.0,>=2.1 in /opt/conda/lib/python3.7/site-packages (from botocore<1.25.0,>=1.24.42->boto3>=1.20.21->sagemaker) (2.8.1)
Requirement already satisfied: zipp>=0.5 in /opt/conda/lib/python3.7/site-packages (from importlib-metadata>=1.4.0->sagemaker) (2.2.0)
Requirement already satisfied: six in /opt/conda/lib/python3.7/site-packages (from packaging>=20.0->sagemaker) (1.14.0)
Requirement already satisfied: pyparsing>=2.0.2 in /opt/conda/lib/python3.7/site-packages (from packaging>=20.0->sagemaker) (2.4.6)
Requirement already satisfied: pytz>=2017.2 in /opt/conda/lib/python3.7/site-packages (from pandas->sagemaker) (2019.3)
Requirement already satisfied: ppft>=1.6.6.4 in /opt/conda/lib/python3.7/site-packages (from pathos->sagemaker) (1.6.6.4)
Requirement already satisfied: pox>=0.3.0 in /opt/conda/lib/python3.7/site-packages (from pathos->sagemaker) (0.3.0)
Requirement already satisfied: dill>=0.3.4 in /opt/conda/lib/python3.7/site-packages (from pathos->sagemaker) (0.3.4)
Requirement already satisfied: multiprocess>=0.70.12 in /opt/conda/lib/python3.7/site-packages (from pathos->sagemaker) (0.70.12.2)
Building wheels for collected packages: sagemaker
Building wheel for sagemaker (setup.py) ... - \ | / - \ | / done
Created wheel for sagemaker: filename=sagemaker-2.86.2-py2.py3-none-any.whl size=720870 sha256=30eb3c1cca0f6a5480b9654912669d9e879aee733c91bd63d8a9addf75998a4d
Stored in directory: /root/.cache/pip/wheels/ea/94/35/1b815c5bdf63f8947160a8e6a4eb12f4bb05bd6f9cc773176a
Successfully built sagemaker
Installing collected packages: botocore, boto3, attrs, sagemaker
Attempting uninstall: botocore
Found existing installation: botocore 1.23.7
Uninstalling botocore-1.23.7:
Successfully uninstalled botocore-1.23.7
Attempting uninstall: boto3
Found existing installation: boto3 1.20.7
Uninstalling boto3-1.20.7:
Successfully uninstalled boto3-1.20.7
Attempting uninstall: attrs
Found existing installation: attrs 19.3.0
Uninstalling attrs-19.3.0:
Successfully uninstalled attrs-19.3.0
Attempting uninstall: sagemaker
Found existing installation: sagemaker 2.69.1.dev0
Uninstalling sagemaker-2.69.1.dev0:
Successfully uninstalled sagemaker-2.69.1.dev0
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
pytest-astropy 0.8.0 requires pytest-cov>=2.0, which is not installed.
pytest-astropy 0.8.0 requires pytest-filter-subpackage>=0.1, which is not installed.
awscli 1.22.7 requires botocore==1.23.7, but you have botocore 1.24.42 which is incompatible.
aiobotocore 1.3.3 requires botocore<1.20.107,>=1.20.106, but you have botocore 1.24.42 which is incompatible.
Successfully installed attrs-20.3.0 boto3-1.21.42 botocore-1.24.42 sagemaker-2.86.2
WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
WARNING: You are using pip version 21.1.3; however, version 22.0.4 is available.
You should consider upgrading via the '/opt/conda/bin/python -m pip install --upgrade pip' command.
Now, upload the data into the default bucket. You can select our own data set for the input_data_uri
as is appropriate.
[3]:
!mkdir -p data
[4]:
local_path = "data/abalone-dataset.csv"
s3 = boto3.resource("s3")
s3.Bucket(f"sagemaker-sample-files").download_file(
"datasets/tabular/uci_abalone/abalone.csv", local_path
)
base_uri = f"s3://{default_bucket}/abalone"
input_data_uri = sagemaker.s3.S3Uploader.upload(
local_path=local_path,
desired_s3_uri=base_uri,
)
print(input_data_uri)
s3://sagemaker-us-west-2-000000000000/abalone/abalone-dataset.csv
Download a second dataset for batch transformation after model creation. You can select our own dataset for the batch_data_uri
as is appropriate.
[5]:
local_path = "data/abalone-dataset-batch"
s3 = boto3.resource("s3")
s3.Bucket(f"sagemaker-servicecatalog-seedcode-{region}").download_file(
"dataset/abalone-dataset-batch", local_path
)
base_uri = f"s3://{default_bucket}/abalone"
batch_data_uri = sagemaker.s3.S3Uploader.upload(
local_path=local_path,
desired_s3_uri=base_uri,
)
print(batch_data_uri)
s3://sagemaker-us-west-2-000000000000/abalone/abalone-dataset-batch
Define Parameters to Parametrize Pipeline Execution
Define Pipeline parameters that you can use to parametrize the pipeline. Parameters enable custom pipeline executions and schedules without having to modify the Pipeline definition.
The supported parameter types include:
ParameterString
- represents astr
Python typeParameterInteger
- represents anint
Python typeParameterFloat
- represents afloat
Python type
These parameters support providing a default value, which can be overridden on pipeline execution. The default value specified should be an instance of the type of the parameter.
The parameters defined in this workflow include:
processing_instance_type
- Theml.*
instance type of the processing job.processing_instance_count
- The instance count of the processing job.instance_type
- Theml.*
instance type of the training job.model_approval_status
- The approval status to register with the trained model for CI/CD purposes (“PendingManualApproval” is the default).input_data
- The S3 bucket URI location of the input data.batch_data
- The S3 bucket URI location of the batch data.mse_threshold
- The Mean Squared Error (MSE) threshold used to verify the accuracy of a model.
[6]:
from sagemaker.workflow.parameters import (
ParameterInteger,
ParameterString,
ParameterFloat,
)
processing_instance_count = ParameterInteger(name="ProcessingInstanceCount", default_value=1)
processing_instance_type = ParameterString(
name="ProcessingInstanceType", default_value="ml.m5.xlarge"
)
instance_type = ParameterString(name="TrainingInstanceType", default_value="ml.m5.xlarge")
model_approval_status = ParameterString(
name="ModelApprovalStatus", default_value="PendingManualApproval"
)
input_data = ParameterString(
name="InputData",
default_value=input_data_uri,
)
batch_data = ParameterString(
name="BatchData",
default_value=batch_data_uri,
)
mse_threshold = ParameterFloat(name="MseThreshold", default_value=6.0)
Define a Processing Step for Feature Engineering
First, develop a preprocessing script that is specified in the Processing step.
This notebook cell writes a file preprocessing_abalone.py
, which contains the preprocessing script. You can update the script, and rerun this cell to overwrite. The preprocessing script uses scikit-learn
to do the following:
Fill in missing sex category data and encode it so that it is suitable for training.
Scale and normalize all numerical fields, aside from sex and rings numerical data.
Split the data into training, validation, and test datasets.
The Processing step executes the script on the input data. The Training step uses the preprocessed training features and labels to train a model. The Evaluation step uses the trained model and preprocessed test features and labels to evaluate the model.
[7]:
!mkdir -p abalone
[8]:
%%writefile abalone/preprocessing.py
import argparse
import os
import requests
import tempfile
import numpy as np
import pandas as pd
from sklearn.compose import ColumnTransformer
from sklearn.impute import SimpleImputer
from sklearn.pipeline import Pipeline
from sklearn.preprocessing import StandardScaler, OneHotEncoder
# Since we get a headerless CSV file, we specify the column names here.
feature_columns_names = [
"sex",
"length",
"diameter",
"height",
"whole_weight",
"shucked_weight",
"viscera_weight",
"shell_weight",
]
label_column = "rings"
feature_columns_dtype = {
"sex": str,
"length": np.float64,
"diameter": np.float64,
"height": np.float64,
"whole_weight": np.float64,
"shucked_weight": np.float64,
"viscera_weight": np.float64,
"shell_weight": np.float64,
}
label_column_dtype = {"rings": np.float64}
def merge_two_dicts(x, y):
z = x.copy()
z.update(y)
return z
if __name__ == "__main__":
base_dir = "/opt/ml/processing"
df = pd.read_csv(
f"{base_dir}/input/abalone-dataset.csv",
header=None,
names=feature_columns_names + [label_column],
dtype=merge_two_dicts(feature_columns_dtype, label_column_dtype),
)
numeric_features = list(feature_columns_names)
numeric_features.remove("sex")
numeric_transformer = Pipeline(
steps=[("imputer", SimpleImputer(strategy="median")), ("scaler", StandardScaler())]
)
categorical_features = ["sex"]
categorical_transformer = Pipeline(
steps=[
("imputer", SimpleImputer(strategy="constant", fill_value="missing")),
("onehot", OneHotEncoder(handle_unknown="ignore")),
]
)
preprocess = ColumnTransformer(
transformers=[
("num", numeric_transformer, numeric_features),
("cat", categorical_transformer, categorical_features),
]
)
y = df.pop("rings")
X_pre = preprocess.fit_transform(df)
y_pre = y.to_numpy().reshape(len(y), 1)
X = np.concatenate((y_pre, X_pre), axis=1)
np.random.shuffle(X)
train, validation, test = np.split(X, [int(0.7 * len(X)), int(0.85 * len(X))])
pd.DataFrame(train).to_csv(f"{base_dir}/train/train.csv", header=False, index=False)
pd.DataFrame(validation).to_csv(
f"{base_dir}/validation/validation.csv", header=False, index=False
)
pd.DataFrame(test).to_csv(f"{base_dir}/test/test.csv", header=False, index=False)
Writing abalone/preprocessing.py
Next, create an instance of a SKLearnProcessor
processor and use that in our ProcessingStep
.
You also specify the framework_version
to use throughout this notebook.
Note the processing_instance_type
and processing_instance_count
parameters used by the processor instance.
[9]:
from sagemaker.sklearn.processing import SKLearnProcessor
framework_version = "1.0-1"
sklearn_processor = SKLearnProcessor(
framework_version=framework_version,
instance_type=processing_instance_type,
instance_count=processing_instance_count,
base_job_name="sklearn-abalone-process",
role=role,
)
Finally, use the processor instance to construct a ProcessingStep
, along with the input and output channels, and the code that runs when the pipeline invokes pipeline execution. This is similar to a processor instance’s run()
method in the Python SDK.
Note the input_data
parameters passed into ProcessingStep
is the input data used in the step. This input data is used by the processor instance when it is run.
Also, note the "train_data"
and "test_data"
named channels specified in the output configuration for the processing job. Step Properties
can be used in subsequent steps and resolve to their runtime values at execution. Specifically, this usage is called out when you define the training step.
[10]:
from sagemaker.processing import ProcessingInput, ProcessingOutput
from sagemaker.workflow.steps import ProcessingStep
step_process = ProcessingStep(
name="AbaloneProcess",
processor=sklearn_processor,
inputs=[
ProcessingInput(source=input_data, destination="/opt/ml/processing/input"),
],
outputs=[
ProcessingOutput(output_name="train", source="/opt/ml/processing/train"),
ProcessingOutput(output_name="validation", source="/opt/ml/processing/validation"),
ProcessingOutput(output_name="test", source="/opt/ml/processing/test"),
],
code="abalone/preprocessing.py",
)
Define a Training Step to Train a Model
In this section, use Amazon SageMaker’s XGBoost Algorithm to train on this dataset. Configure an Estimator for the XGBoost algorithm and the input dataset. A typical training script loads data from the input channels, configures training with hyperparameters, trains a model, and saves a model to model_dir
so that it can be hosted later.
The model path where the models from training are saved is also specified.
Note the instance_type
parameter may be used in multiple places in the pipeline. In this case, the instance_type
is passed into the estimator.
[11]:
from sagemaker.estimator import Estimator
model_path = f"s3://{default_bucket}/AbaloneTrain"
image_uri = sagemaker.image_uris.retrieve(
framework="xgboost",
region=region,
version="1.0-1",
py_version="py3",
instance_type=instance_type,
)
xgb_train = Estimator(
image_uri=image_uri,
instance_type=instance_type,
instance_count=1,
output_path=model_path,
role=role,
)
xgb_train.set_hyperparameters(
objective="reg:linear",
num_round=50,
max_depth=5,
eta=0.2,
gamma=4,
min_child_weight=6,
subsample=0.7,
)
Finally, use the estimator instance to construct a TrainingStep
as well as the properties
of the prior ProcessingStep
used as input in the TrainingStep
inputs and the code that’s executed when the pipeline invokes the pipeline execution. This is similar to an estimator’s fit
method in the Python SDK.
Pass in the S3Uri
of the "train_data"
output channel to the TrainingStep
. Also, use the other "test_data"
output channel for model evaluation in the pipeline. The properties
attribute of a Pipeline step matches the object model of the corresponding response of a describe call. These properties can be referenced as placeholder values and are resolved at runtime. For example, the ProcessingStep
properties
attribute matches the object model of the
DescribeProcessingJob response object.
[12]:
from sagemaker.inputs import TrainingInput
from sagemaker.workflow.steps import TrainingStep
step_train = TrainingStep(
name="AbaloneTrain",
estimator=xgb_train,
inputs={
"train": TrainingInput(
s3_data=step_process.properties.ProcessingOutputConfig.Outputs["train"].S3Output.S3Uri,
content_type="text/csv",
),
"validation": TrainingInput(
s3_data=step_process.properties.ProcessingOutputConfig.Outputs[
"validation"
].S3Output.S3Uri,
content_type="text/csv",
),
},
)
Define a Model Evaluation Step to Evaluate the Trained Model
First, develop an evaluation script that is specified in a Processing step that performs the model evaluation.
After pipeline execution, you can examine the resulting evaluation.json
for analysis.
The evaluation script uses xgboost
to do the following:
Load the model.
Read the test data.
Issue predictions against the test data.
Build a classification report, including accuracy and ROC curve.
Save the evaluation report to the evaluation directory.
[13]:
%%writefile abalone/evaluation.py
import json
import pathlib
import pickle
import tarfile
import joblib
import numpy as np
import pandas as pd
import xgboost
from sklearn.metrics import mean_squared_error
if __name__ == "__main__":
model_path = f"/opt/ml/processing/model/model.tar.gz"
with tarfile.open(model_path) as tar:
tar.extractall(path=".")
model = pickle.load(open("xgboost-model", "rb"))
test_path = "/opt/ml/processing/test/test.csv"
df = pd.read_csv(test_path, header=None)
y_test = df.iloc[:, 0].to_numpy()
df.drop(df.columns[0], axis=1, inplace=True)
X_test = xgboost.DMatrix(df.values)
predictions = model.predict(X_test)
mse = mean_squared_error(y_test, predictions)
std = np.std(y_test - predictions)
report_dict = {
"regression_metrics": {
"mse": {"value": mse, "standard_deviation": std},
},
}
output_dir = "/opt/ml/processing/evaluation"
pathlib.Path(output_dir).mkdir(parents=True, exist_ok=True)
evaluation_path = f"{output_dir}/evaluation.json"
with open(evaluation_path, "w") as f:
f.write(json.dumps(report_dict))
Writing abalone/evaluation.py
Next, create an instance of a ScriptProcessor
processor and use it in the ProcessingStep
.
Note the processing_instance_type
parameter passed into the processor.
[14]:
from sagemaker.processing import ScriptProcessor
script_eval = ScriptProcessor(
image_uri=image_uri,
command=["python3"],
instance_type=processing_instance_type,
instance_count=1,
base_job_name="script-abalone-eval",
role=role,
)
Use the processor instance to construct a ProcessingStep
, along with the input and output channels and the code that runs when the pipeline invokes pipeline execution. This is similar to a processor instance’s run
method in the Python SDK.
Specifically, the S3ModelArtifacts
from the step_train
properties
and the S3Uri
of the "test_data"
output channel of the step_process
properties
are passed as inputs. The TrainingStep
and ProcessingStep
properties
attribute matches the object model of the DescribeTrainingJob and
DescribeProcessingJob response objects, respectively.
[15]:
from sagemaker.workflow.properties import PropertyFile
evaluation_report = PropertyFile(
name="EvaluationReport", output_name="evaluation", path="evaluation.json"
)
step_eval = ProcessingStep(
name="AbaloneEval",
processor=script_eval,
inputs=[
ProcessingInput(
source=step_train.properties.ModelArtifacts.S3ModelArtifacts,
destination="/opt/ml/processing/model",
),
ProcessingInput(
source=step_process.properties.ProcessingOutputConfig.Outputs["test"].S3Output.S3Uri,
destination="/opt/ml/processing/test",
),
],
outputs=[
ProcessingOutput(output_name="evaluation", source="/opt/ml/processing/evaluation"),
],
code="abalone/evaluation.py",
property_files=[evaluation_report],
)
Define a Create Model Step to Create a Model
In order to perform batch transformation using the example model, create a SageMaker model.
Specifically, pass in the S3ModelArtifacts
from the TrainingStep
, step_train
properties. The TrainingStep
properties
attribute matches the object model of the DescribeTrainingJob response object.
[16]:
from sagemaker.model import Model
model = Model(
image_uri=image_uri,
model_data=step_train.properties.ModelArtifacts.S3ModelArtifacts,
sagemaker_session=sagemaker_session,
role=role,
)
Supply the model input (instance_type
and accelerator_type
) for creating the SageMaker Model, and then define the CreateModelStep
, passing in the inputs and the model instance defined before.
[17]:
from sagemaker.inputs import CreateModelInput
from sagemaker.workflow.steps import CreateModelStep
inputs = CreateModelInput(
instance_type="ml.m5.large",
accelerator_type="ml.eia1.medium",
)
step_create_model = CreateModelStep(
name="AbaloneCreateModel",
model=model,
inputs=inputs,
)
Define a Transform Step to Perform Batch Transformation
Now that a model instance is defined, create a Transformer
instance with the appropriate model type, compute instance type, and desired output S3 URI.
Specifically, pass in the ModelName
from the CreateModelStep
, step_create_model
properties. The CreateModelStep
properties
attribute matches the object model of the DescribeModel response object.
[18]:
from sagemaker.transformer import Transformer
transformer = Transformer(
model_name=step_create_model.properties.ModelName,
instance_type="ml.m5.xlarge",
instance_count=1,
output_path=f"s3://{default_bucket}/AbaloneTransform",
)
Pass in the transformer instance and the TransformInput
with the batch_data
pipeline parameter defined earlier.
[19]:
from sagemaker.inputs import TransformInput
from sagemaker.workflow.steps import TransformStep
step_transform = TransformStep(
name="AbaloneTransform", transformer=transformer, inputs=TransformInput(data=batch_data)
)
Define a Register Model Step to Create a Model Package
Use the estimator instance specified in the training step to construct an instance of RegisterModel
. The result of executing RegisterModel
in a pipeline is a model package. A model package is an abstraction of reusable model artifacts that packages all ingredients required for inference. Primarily, it consists of an inference specification that defines the inference image to use along with an optional model weights location.
A model package group is a collection of model packages. A model package group can be created for a specific ML business problem, and new versions of the model packages can be added to it. Typically, customers are expected to create a ModelPackageGroup for a SageMaker pipeline so that model package versions can be added to the group for every SageMaker Pipeline run.
The construction of RegisterModel
is similar to an estimator instance’s register
method in the Python SDK.
Specifically, pass in the S3ModelArtifacts
from the TrainingStep
, step_train
properties. The TrainingStep
properties
attribute matches the object model of the DescribeTrainingJob response object.
Note that the specific model package group name provided in this notebook can be used in the model registry and CI/CD work with SageMaker Projects.
[20]:
from sagemaker.model_metrics import MetricsSource, ModelMetrics
from sagemaker.workflow.step_collections import RegisterModel
model_metrics = ModelMetrics(
model_statistics=MetricsSource(
s3_uri="{}/evaluation.json".format(
step_eval.arguments["ProcessingOutputConfig"]["Outputs"][0]["S3Output"]["S3Uri"]
),
content_type="application/json",
)
)
step_register = RegisterModel(
name="AbaloneRegisterModel",
estimator=xgb_train,
model_data=step_train.properties.ModelArtifacts.S3ModelArtifacts,
content_types=["text/csv"],
response_types=["text/csv"],
inference_instances=["ml.t2.medium", "ml.m5.xlarge"],
transform_instances=["ml.m5.xlarge"],
model_package_group_name=model_package_group_name,
approval_status=model_approval_status,
model_metrics=model_metrics,
)
Define a Fail Step to Terminate the Pipeline Execution and Mark it as Failed
This section walks you through the following steps:
Define a
FailStep
with customized error message, which indicates the cause of the execution failure.Enter the
FailStep
error message with aJoin
function, which appends a static text string with the dynamicmse_threshold
parameter to build a more informative error message.
[21]:
from sagemaker.workflow.fail_step import FailStep
from sagemaker.workflow.functions import Join
step_fail = FailStep(
name="AbaloneMSEFail",
error_message=Join(on=" ", values=["Execution failed due to MSE >", mse_threshold]),
)
Define a Condition Step to Check Accuracy and Conditionally Create a Model and Run a Batch Transformation and Register a Model in the Model Registry, Or Terminate the Execution in Failed State
In this step, the model is registered only if the accuracy of the model, as determined by the evaluation step step_eval
, exceeded a specified value. Otherwise, the pipeline execution fails and terminates. A ConditionStep
enables pipelines to support conditional execution in the pipeline DAG based on the conditions of the step properties.
In the following section, you:
Define a
ConditionLessThanOrEqualTo
on the accuracy value found in the output of the evaluation step,step_eval
.Use the condition in the list of conditions in a
ConditionStep
.Pass the
CreateModelStep
andTransformStep
steps, and theRegisterModel
step collection into theif_steps
of theConditionStep
, which are only executed if the condition evaluates toTrue
.Pass the
FailStep
step into theelse_steps
of theConditionStep
, which is only executed if the condition evaluates toFalse
.
[22]:
from sagemaker.workflow.conditions import ConditionLessThanOrEqualTo
from sagemaker.workflow.condition_step import ConditionStep
from sagemaker.workflow.functions import JsonGet
cond_lte = ConditionLessThanOrEqualTo(
left=JsonGet(
step_name=step_eval.name,
property_file=evaluation_report,
json_path="regression_metrics.mse.value",
),
right=mse_threshold,
)
step_cond = ConditionStep(
name="AbaloneMSECond",
conditions=[cond_lte],
if_steps=[step_register, step_create_model, step_transform],
else_steps=[step_fail],
)
Define a Pipeline of Parameters, Steps, and Conditions
In this section, combine the steps into a Pipeline so it can be executed.
A pipeline requires a name
, parameters
, and steps
. Names must be unique within an (account, region)
pair.
Note:
All the parameters used in the definitions must be present.
Steps passed into the pipeline do not have to be listed in the order of execution. The SageMaker Pipeline service resolves the data dependency DAG as steps for the execution to complete.
Steps must be unique to across the pipeline step list and all condition step if/else lists.
[23]:
from sagemaker.workflow.pipeline import Pipeline
pipeline_name = f"AbalonePipeline"
pipeline = Pipeline(
name=pipeline_name,
parameters=[
processing_instance_type,
processing_instance_count,
instance_type,
model_approval_status,
input_data,
batch_data,
mse_threshold,
],
steps=[step_process, step_train, step_eval, step_cond],
)
(Optional) Examining the pipeline definition
The JSON of the pipeline definition can be examined to confirm the pipeline is well-defined and the parameters and step properties resolve correctly.
[24]:
import json
definition = json.loads(pipeline.definition())
definition
No finished training job found associated with this estimator. Please make sure this estimator is only used for building workflow config
[24]:
{'Version': '2020-12-01',
'Metadata': {},
'Parameters': [{'Name': 'ProcessingInstanceType',
'Type': 'String',
'DefaultValue': 'ml.m5.xlarge'},
{'Name': 'ProcessingInstanceCount', 'Type': 'Integer', 'DefaultValue': 1},
{'Name': 'TrainingInstanceType',
'Type': 'String',
'DefaultValue': 'ml.m5.xlarge'},
{'Name': 'ModelApprovalStatus',
'Type': 'String',
'DefaultValue': 'PendingManualApproval'},
{'Name': 'InputData',
'Type': 'String',
'DefaultValue': 's3://sagemaker-us-west-2-000000000000/abalone/abalone-dataset.csv'},
{'Name': 'BatchData',
'Type': 'String',
'DefaultValue': 's3://sagemaker-us-west-2-000000000000/abalone/abalone-dataset-batch'},
{'Name': 'MseThreshold', 'Type': 'Float', 'DefaultValue': 6.0}],
'PipelineExperimentConfig': {'ExperimentName': {'Get': 'Execution.PipelineName'},
'TrialName': {'Get': 'Execution.PipelineExecutionId'}},
'Steps': [{'Name': 'AbaloneProcess',
'Type': 'Processing',
'Arguments': {'ProcessingResources': {'ClusterConfig': {'InstanceType': {'Get': 'Parameters.ProcessingInstanceType'},
'InstanceCount': {'Get': 'Parameters.ProcessingInstanceCount'},
'VolumeSizeInGB': 30}},
'AppSpecification': {'ImageUri': '246618743249.dkr.ecr.us-west-2.amazonaws.com/sagemaker-scikit-learn:1.0-1-cpu-py3',
'ContainerEntrypoint': ['python3',
'/opt/ml/processing/input/code/preprocessing.py']},
'RoleArn': 'arn:aws:iam::000000000000:role/ProdBuildSystemStack-ReleaseBuildRoleFB326D49-QK8LUA2UI1IC',
'ProcessingInputs': [{'InputName': 'input-1',
'AppManaged': False,
'S3Input': {'S3Uri': {'Get': 'Parameters.InputData'},
'LocalPath': '/opt/ml/processing/input',
'S3DataType': 'S3Prefix',
'S3InputMode': 'File',
'S3DataDistributionType': 'FullyReplicated',
'S3CompressionType': 'None'}},
{'InputName': 'code',
'AppManaged': False,
'S3Input': {'S3Uri': 's3://sagemaker-us-west-2-000000000000/AbaloneProcess-a55d50a0c87783b887401dc1ff1d9bf9/input/code/preprocessing.py',
'LocalPath': '/opt/ml/processing/input/code',
'S3DataType': 'S3Prefix',
'S3InputMode': 'File',
'S3DataDistributionType': 'FullyReplicated',
'S3CompressionType': 'None'}}],
'ProcessingOutputConfig': {'Outputs': [{'OutputName': 'train',
'AppManaged': False,
'S3Output': {'S3Uri': 's3://sagemaker-us-west-2-000000000000/AbaloneProcess-a55d50a0c87783b887401dc1ff1d9bf9/output/train',
'LocalPath': '/opt/ml/processing/train',
'S3UploadMode': 'EndOfJob'}},
{'OutputName': 'validation',
'AppManaged': False,
'S3Output': {'S3Uri': 's3://sagemaker-us-west-2-000000000000/AbaloneProcess-a55d50a0c87783b887401dc1ff1d9bf9/output/validation',
'LocalPath': '/opt/ml/processing/validation',
'S3UploadMode': 'EndOfJob'}},
{'OutputName': 'test',
'AppManaged': False,
'S3Output': {'S3Uri': 's3://sagemaker-us-west-2-000000000000/AbaloneProcess-a55d50a0c87783b887401dc1ff1d9bf9/output/test',
'LocalPath': '/opt/ml/processing/test',
'S3UploadMode': 'EndOfJob'}}]}}},
{'Name': 'AbaloneTrain',
'Type': 'Training',
'Arguments': {'AlgorithmSpecification': {'TrainingInputMode': 'File',
'TrainingImage': '246618743249.dkr.ecr.us-west-2.amazonaws.com/sagemaker-xgboost:1.0-1-cpu-py3'},
'OutputDataConfig': {'S3OutputPath': 's3://sagemaker-us-west-2-000000000000/AbaloneTrain'},
'StoppingCondition': {'MaxRuntimeInSeconds': 86400},
'ResourceConfig': {'InstanceCount': 1,
'InstanceType': {'Get': 'Parameters.TrainingInstanceType'},
'VolumeSizeInGB': 30},
'RoleArn': 'arn:aws:iam::000000000000:role/ProdBuildSystemStack-ReleaseBuildRoleFB326D49-QK8LUA2UI1IC',
'InputDataConfig': [{'DataSource': {'S3DataSource': {'S3DataType': 'S3Prefix',
'S3Uri': {'Get': "Steps.AbaloneProcess.ProcessingOutputConfig.Outputs['train'].S3Output.S3Uri"},
'S3DataDistributionType': 'FullyReplicated'}},
'ContentType': 'text/csv',
'ChannelName': 'train'},
{'DataSource': {'S3DataSource': {'S3DataType': 'S3Prefix',
'S3Uri': {'Get': "Steps.AbaloneProcess.ProcessingOutputConfig.Outputs['validation'].S3Output.S3Uri"},
'S3DataDistributionType': 'FullyReplicated'}},
'ContentType': 'text/csv',
'ChannelName': 'validation'}],
'HyperParameters': {'objective': 'reg:linear',
'num_round': '50',
'max_depth': '5',
'eta': '0.2',
'gamma': '4',
'min_child_weight': '6',
'subsample': '0.7'},
'ProfilerRuleConfigurations': [{'RuleConfigurationName': 'ProfilerReport-1650241632',
'RuleEvaluatorImage': '895741380848.dkr.ecr.us-west-2.amazonaws.com/sagemaker-debugger-rules:latest',
'RuleParameters': {'rule_to_invoke': 'ProfilerReport'}}],
'ProfilerConfig': {'S3OutputPath': 's3://sagemaker-us-west-2-000000000000/AbaloneTrain'}}},
{'Name': 'AbaloneEval',
'Type': 'Processing',
'Arguments': {'ProcessingResources': {'ClusterConfig': {'InstanceType': {'Get': 'Parameters.ProcessingInstanceType'},
'InstanceCount': 1,
'VolumeSizeInGB': 30}},
'AppSpecification': {'ImageUri': '246618743249.dkr.ecr.us-west-2.amazonaws.com/sagemaker-xgboost:1.0-1-cpu-py3',
'ContainerEntrypoint': ['python3',
'/opt/ml/processing/input/code/evaluation.py']},
'RoleArn': 'arn:aws:iam::000000000000:role/ProdBuildSystemStack-ReleaseBuildRoleFB326D49-QK8LUA2UI1IC',
'ProcessingInputs': [{'InputName': 'input-1',
'AppManaged': False,
'S3Input': {'S3Uri': {'Get': 'Steps.AbaloneTrain.ModelArtifacts.S3ModelArtifacts'},
'LocalPath': '/opt/ml/processing/model',
'S3DataType': 'S3Prefix',
'S3InputMode': 'File',
'S3DataDistributionType': 'FullyReplicated',
'S3CompressionType': 'None'}},
{'InputName': 'input-2',
'AppManaged': False,
'S3Input': {'S3Uri': {'Get': "Steps.AbaloneProcess.ProcessingOutputConfig.Outputs['test'].S3Output.S3Uri"},
'LocalPath': '/opt/ml/processing/test',
'S3DataType': 'S3Prefix',
'S3InputMode': 'File',
'S3DataDistributionType': 'FullyReplicated',
'S3CompressionType': 'None'}},
{'InputName': 'code',
'AppManaged': False,
'S3Input': {'S3Uri': 's3://sagemaker-us-west-2-000000000000/AbaloneEval-3419a950c93b74404394e09ce2edc51e/input/code/evaluation.py',
'LocalPath': '/opt/ml/processing/input/code',
'S3DataType': 'S3Prefix',
'S3InputMode': 'File',
'S3DataDistributionType': 'FullyReplicated',
'S3CompressionType': 'None'}}],
'ProcessingOutputConfig': {'Outputs': [{'OutputName': 'evaluation',
'AppManaged': False,
'S3Output': {'S3Uri': 's3://sagemaker-us-west-2-000000000000/AbaloneEval-3419a950c93b74404394e09ce2edc51e/output/evaluation',
'LocalPath': '/opt/ml/processing/evaluation',
'S3UploadMode': 'EndOfJob'}}]}},
'PropertyFiles': [{'PropertyFileName': 'EvaluationReport',
'OutputName': 'evaluation',
'FilePath': 'evaluation.json'}]},
{'Name': 'AbaloneMSECond',
'Type': 'Condition',
'Arguments': {'Conditions': [{'Type': 'LessThanOrEqualTo',
'LeftValue': {'Std:JsonGet': {'PropertyFile': {'Get': 'Steps.AbaloneEval.PropertyFiles.EvaluationReport'},
'Path': 'regression_metrics.mse.value'}},
'RightValue': {'Get': 'Parameters.MseThreshold'}}],
'IfSteps': [{'Name': 'AbaloneRegisterModel',
'Type': 'RegisterModel',
'Arguments': {'ModelPackageGroupName': 'AbaloneModelPackageGroupName',
'ModelMetrics': {'ModelQuality': {'Statistics': {'ContentType': 'application/json',
'S3Uri': 's3://sagemaker-us-west-2-000000000000/AbaloneEval-3419a950c93b74404394e09ce2edc51e/output/evaluation/evaluation.json'}},
'Bias': {},
'Explainability': {}},
'InferenceSpecification': {'Containers': [{'Image': '246618743249.dkr.ecr.us-west-2.amazonaws.com/sagemaker-xgboost:1.0-1-cpu-py3',
'ModelDataUrl': {'Get': 'Steps.AbaloneTrain.ModelArtifacts.S3ModelArtifacts'}}],
'SupportedContentTypes': ['text/csv'],
'SupportedResponseMIMETypes': ['text/csv'],
'SupportedRealtimeInferenceInstanceTypes': ['ml.t2.medium',
'ml.m5.xlarge'],
'SupportedTransformInstanceTypes': ['ml.m5.xlarge']},
'ModelApprovalStatus': {'Get': 'Parameters.ModelApprovalStatus'}}},
{'Name': 'AbaloneCreateModel',
'Type': 'Model',
'Arguments': {'ExecutionRoleArn': 'arn:aws:iam::000000000000:role/ProdBuildSystemStack-ReleaseBuildRoleFB326D49-QK8LUA2UI1IC',
'PrimaryContainer': {'Image': '246618743249.dkr.ecr.us-west-2.amazonaws.com/sagemaker-xgboost:1.0-1-cpu-py3',
'Environment': {},
'ModelDataUrl': {'Get': 'Steps.AbaloneTrain.ModelArtifacts.S3ModelArtifacts'}}}},
{'Name': 'AbaloneTransform',
'Type': 'Transform',
'Arguments': {'ModelName': {'Get': 'Steps.AbaloneCreateModel.ModelName'},
'TransformInput': {'DataSource': {'S3DataSource': {'S3DataType': 'S3Prefix',
'S3Uri': {'Get': 'Parameters.BatchData'}}}},
'TransformOutput': {'S3OutputPath': 's3://sagemaker-us-west-2-000000000000/AbaloneTransform'},
'TransformResources': {'InstanceCount': 1,
'InstanceType': 'ml.m5.xlarge'}}}],
'ElseSteps': [{'Name': 'AbaloneMSEFail',
'Type': 'Fail',
'Arguments': {'ErrorMessage': {'Std:Join': {'On': ' ',
'Values': ['Execution failed due to MSE >',
{'Get': 'Parameters.MseThreshold'}]}}}}]}}]}
Submit the pipeline to SageMaker and start execution
Submit the pipeline definition to the Pipeline service. The Pipeline service uses the role that is passed in to create all the jobs defined in the steps.
[25]:
pipeline.upsert(role_arn=role)
No finished training job found associated with this estimator. Please make sure this estimator is only used for building workflow config
No finished training job found associated with this estimator. Please make sure this estimator is only used for building workflow config
[25]:
{'PipelineArn': 'arn:aws:sagemaker:us-west-2:000000000000:pipeline/abalonepipeline',
'ResponseMetadata': {'RequestId': '0dafe696-f7b9-497d-883f-46dd6db4d900',
'HTTPStatusCode': 200,
'HTTPHeaders': {'x-amzn-requestid': '0dafe696-f7b9-497d-883f-46dd6db4d900',
'content-type': 'application/x-amz-json-1.1',
'content-length': '83',
'date': 'Mon, 18 Apr 2022 00:27:13 GMT'},
'RetryAttempts': 0}}
Start the pipeline and accept all the default parameters.
[26]:
execution = pipeline.start()
Pipeline Operations: Examining and Waiting for Pipeline Execution
Describe the pipeline execution.
[27]:
execution.describe()
[27]:
{'PipelineArn': 'arn:aws:sagemaker:us-west-2:000000000000:pipeline/abalonepipeline',
'PipelineExecutionArn': 'arn:aws:sagemaker:us-west-2:000000000000:pipeline/abalonepipeline/execution/3a9dqd5mmh4u',
'PipelineExecutionDisplayName': 'execution-1650241635185',
'PipelineExecutionStatus': 'Executing',
'CreationTime': datetime.datetime(2022, 4, 18, 0, 27, 15, 69000, tzinfo=tzlocal()),
'LastModifiedTime': datetime.datetime(2022, 4, 18, 0, 27, 15, 69000, tzinfo=tzlocal()),
'CreatedBy': {},
'LastModifiedBy': {},
'ResponseMetadata': {'RequestId': '59d80668-c513-47e4-b703-d12d838faafb',
'HTTPStatusCode': 200,
'HTTPHeaders': {'x-amzn-requestid': '59d80668-c513-47e4-b703-d12d838faafb',
'content-type': 'application/x-amz-json-1.1',
'content-length': '395',
'date': 'Mon, 18 Apr 2022 00:27:14 GMT'},
'RetryAttempts': 0}}
Wait for the execution to complete.
[28]:
execution.wait()
List the steps in the execution. These are the steps in the pipeline that have been resolved by the step executor service.
[29]:
execution.list_steps()
[29]:
[{'StepName': 'AbaloneTransform',
'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 21, 710000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 43, 5, 175000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'TransformJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:transform-job/pipelines-3a9dqd5mmh4u-abalonetransform-shmqcpjkhl'}}},
{'StepName': 'AbaloneCreateModel',
'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 20, 252000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 21, 188000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'Model': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:model/pipelines-3a9dqd5mmh4u-abalonecreatemodel-nr28ooxmjn'}}},
{'StepName': 'AbaloneRegisterModel',
'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 20, 252000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 22, 18000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'RegisterModel': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:model-package/abalonemodelpackagegroupname/691'}}},
{'StepName': 'AbaloneMSECond',
'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 19, 251000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 19, 670000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'Condition': {'Outcome': 'True'}}},
{'StepName': 'AbaloneEval',
'StartTime': datetime.datetime(2022, 4, 18, 0, 34, 22, 219000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 18, 302000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-3a9dqd5mmh4u-abaloneeval-rqz0ph05st'}}},
{'StepName': 'AbaloneTrain',
'StartTime': datetime.datetime(2022, 4, 18, 0, 31, 35, 376000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 34, 21, 590000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'TrainingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:training-job/pipelines-3a9dqd5mmh4u-abalonetrain-slip45cf16'}}},
{'StepName': 'AbaloneProcess',
'StartTime': datetime.datetime(2022, 4, 18, 0, 27, 16, 184000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 31, 34, 404000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-3a9dqd5mmh4u-abaloneprocess-ck1tj9s93u'}}}]
Examining the Evaluation
Examine the resulting model evaluation after the pipeline completes. Download the resulting evaluation.json
file from S3 and print the report.
[30]:
from pprint import pprint
evaluation_json = sagemaker.s3.S3Downloader.read_file(
"{}/evaluation.json".format(
step_eval.arguments["ProcessingOutputConfig"]["Outputs"][0]["S3Output"]["S3Uri"]
)
)
pprint(json.loads(evaluation_json))
{'regression_metrics': {'mse': {'standard_deviation': 2.1549187830947223,
'value': 4.64537577872224}}}
Lineage
Review the lineage of the artifacts generated by the pipeline.
[31]:
import time
from sagemaker.lineage.visualizer import LineageTableVisualizer
viz = LineageTableVisualizer(sagemaker.session.Session())
for execution_step in reversed(execution.list_steps()):
print(execution_step)
display(viz.show(pipeline_execution_step=execution_step))
time.sleep(5)
{'StepName': 'AbaloneProcess', 'StartTime': datetime.datetime(2022, 4, 18, 0, 27, 16, 184000, tzinfo=tzlocal()), 'EndTime': datetime.datetime(2022, 4, 18, 0, 31, 34, 404000, tzinfo=tzlocal()), 'StepStatus': 'Succeeded', 'AttemptCount': 0, 'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-3a9dqd5mmh4u-abaloneprocess-ck1tj9s93u'}}}
Name/Source | Direction | Type | Association Type | Lineage Type | |
---|---|---|---|---|---|
0 | s3://...1dc1ff1d9bf9/input/code/preprocessing.py | Input | DataSet | ContributedTo | artifact |
1 | s3://...000000000000/abalone/abalone-dataset.csv | Input | DataSet | ContributedTo | artifact |
2 | 24661...om/sagemaker-scikit-learn:1.0-1-cpu-py3 | Input | Image | ContributedTo | artifact |
3 | s3://...50a0c87783b887401dc1ff1d9bf9/output/test | Output | DataSet | Produced | artifact |
4 | s3://...7783b887401dc1ff1d9bf9/output/validation | Output | DataSet | Produced | artifact |
5 | s3://...0a0c87783b887401dc1ff1d9bf9/output/train | Output | DataSet | Produced | artifact |
{'StepName': 'AbaloneTrain', 'StartTime': datetime.datetime(2022, 4, 18, 0, 31, 35, 376000, tzinfo=tzlocal()), 'EndTime': datetime.datetime(2022, 4, 18, 0, 34, 21, 590000, tzinfo=tzlocal()), 'StepStatus': 'Succeeded', 'AttemptCount': 0, 'Metadata': {'TrainingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:training-job/pipelines-3a9dqd5mmh4u-abalonetrain-slip45cf16'}}}
Name/Source | Direction | Type | Association Type | Lineage Type | |
---|---|---|---|---|---|
0 | s3://...7783b887401dc1ff1d9bf9/output/validation | Input | DataSet | ContributedTo | artifact |
1 | s3://...0a0c87783b887401dc1ff1d9bf9/output/train | Input | DataSet | ContributedTo | artifact |
2 | 24661...naws.com/sagemaker-xgboost:1.0-1-cpu-py3 | Input | Image | ContributedTo | artifact |
3 | s3://...loneTrain-SLIp45cF16/output/model.tar.gz | Output | Model | Produced | artifact |
{'StepName': 'AbaloneEval', 'StartTime': datetime.datetime(2022, 4, 18, 0, 34, 22, 219000, tzinfo=tzlocal()), 'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 18, 302000, tzinfo=tzlocal()), 'StepStatus': 'Succeeded', 'AttemptCount': 0, 'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-3a9dqd5mmh4u-abaloneeval-rqz0ph05st'}}}
Name/Source | Direction | Type | Association Type | Lineage Type | |
---|---|---|---|---|---|
0 | s3://...394e09ce2edc51e/input/code/evaluation.py | Input | DataSet | ContributedTo | artifact |
1 | s3://...50a0c87783b887401dc1ff1d9bf9/output/test | Input | DataSet | ContributedTo | artifact |
2 | s3://...loneTrain-SLIp45cF16/output/model.tar.gz | Input | Model | ContributedTo | artifact |
3 | 24661...naws.com/sagemaker-xgboost:1.0-1-cpu-py3 | Input | Image | ContributedTo | artifact |
4 | s3://...3b74404394e09ce2edc51e/output/evaluation | Output | DataSet | Produced | artifact |
{'StepName': 'AbaloneMSECond', 'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 19, 251000, tzinfo=tzlocal()), 'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 19, 670000, tzinfo=tzlocal()), 'StepStatus': 'Succeeded', 'AttemptCount': 0, 'Metadata': {'Condition': {'Outcome': 'True'}}}
None
{'StepName': 'AbaloneRegisterModel', 'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 20, 252000, tzinfo=tzlocal()), 'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 22, 18000, tzinfo=tzlocal()), 'StepStatus': 'Succeeded', 'AttemptCount': 0, 'Metadata': {'RegisterModel': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:model-package/abalonemodelpackagegroupname/691'}}}
Name/Source | Direction | Type | Association Type | Lineage Type | |
---|---|---|---|---|---|
0 | s3://...loneTrain-SLIp45cF16/output/model.tar.gz | Input | Model | ContributedTo | artifact |
1 | 24661...naws.com/sagemaker-xgboost:1.0-1-cpu-py3 | Input | Image | ContributedTo | artifact |
2 | abalonemodelpackagegroupname-691-PendingManual... | Input | Approval | ContributedTo | action |
3 | AbaloneModelPackageGroupName-1607556271-aws-mo... | Output | ModelGroup | AssociatedWith | context |
{'StepName': 'AbaloneCreateModel', 'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 20, 252000, tzinfo=tzlocal()), 'EndTime': datetime.datetime(2022, 4, 18, 0, 38, 21, 188000, tzinfo=tzlocal()), 'StepStatus': 'Succeeded', 'AttemptCount': 0, 'Metadata': {'Model': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:model/pipelines-3a9dqd5mmh4u-abalonecreatemodel-nr28ooxmjn'}}}
None
{'StepName': 'AbaloneTransform', 'StartTime': datetime.datetime(2022, 4, 18, 0, 38, 21, 710000, tzinfo=tzlocal()), 'EndTime': datetime.datetime(2022, 4, 18, 0, 43, 5, 175000, tzinfo=tzlocal()), 'StepStatus': 'Succeeded', 'AttemptCount': 0, 'Metadata': {'TransformJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:transform-job/pipelines-3a9dqd5mmh4u-abalonetransform-shmqcpjkhl'}}}
Name/Source | Direction | Type | Association Type | Lineage Type | |
---|---|---|---|---|---|
0 | s3://...loneTrain-SLIp45cF16/output/model.tar.gz | Input | Model | ContributedTo | artifact |
1 | 24661...naws.com/sagemaker-xgboost:1.0-1-cpu-py3 | Input | Image | ContributedTo | artifact |
2 | s3://...1695447989/abalone/abalone-dataset-batch | Input | DataSet | ContributedTo | artifact |
3 | s3://...-us-west-2-000000000000/AbaloneTransform | Output | DataSet | Produced | artifact |
Parametrized Executions
You can run additional executions of the pipeline and specify different pipeline parameters. The parameters
argument is a dictionary containing parameter names, and where the values are used to override the defaults values.
Based on the performance of the model, you might want to kick off another pipeline execution on a compute-optimized instance type and set the model approval status to “Approved” automatically. This means that the model package version generated by the RegisterModel
step is automatically ready for deployment through CI/CD pipelines, such as with SageMaker Projects.
[32]:
execution = pipeline.start(
parameters=dict(
ProcessingInstanceType="ml.c5.xlarge",
ModelApprovalStatus="Approved",
)
)
[33]:
execution.wait()
[34]:
execution.list_steps()
[34]:
[{'StepName': 'AbaloneTransform',
'StartTime': datetime.datetime(2022, 4, 18, 0, 55, 17, 609000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 59, 24, 521000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'TransformJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:transform-job/pipelines-9275502uvwe2-abalonetransform-sdqf868czf'}}},
{'StepName': 'AbaloneCreateModel',
'StartTime': datetime.datetime(2022, 4, 18, 0, 55, 16, 365000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 55, 17, 124000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'Model': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:model/pipelines-9275502uvwe2-abalonecreatemodel-qbuz4uirnh'}}},
{'StepName': 'AbaloneRegisterModel',
'StartTime': datetime.datetime(2022, 4, 18, 0, 55, 16, 365000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 55, 16, 953000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'RegisterModel': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:model-package/abalonemodelpackagegroupname/692'}}},
{'StepName': 'AbaloneMSECond',
'StartTime': datetime.datetime(2022, 4, 18, 0, 55, 15, 608000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 55, 15, 837000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'Condition': {'Outcome': 'True'}}},
{'StepName': 'AbaloneEval',
'StartTime': datetime.datetime(2022, 4, 18, 0, 51, 6, 271000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 55, 14, 745000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-9275502uvwe2-abaloneeval-v43er0jak2'}}},
{'StepName': 'AbaloneTrain',
'StartTime': datetime.datetime(2022, 4, 18, 0, 48, 20, 426000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 51, 5, 564000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'TrainingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:training-job/pipelines-9275502uvwe2-abalonetrain-6rtveegego'}}},
{'StepName': 'AbaloneProcess',
'StartTime': datetime.datetime(2022, 4, 18, 0, 43, 58, 687000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 0, 48, 18, 976000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-9275502uvwe2-abaloneprocess-vynjsz7vps'}}}]
Apart from that, you might also want to adjust the MSE threshold to a smaller value and raise the bar for the accuracy of the registered model. In this case you can override the MSE threshold like the following:
[35]:
execution = pipeline.start(parameters=dict(MseThreshold=3.0))
If the MSE threshold is not satisfied, the pipeline execution enters the FailStep
and is marked as failed.
[36]:
try:
execution.wait()
except Exception as error:
print(error)
Waiter PipelineExecutionComplete failed: Waiter encountered a terminal failure state: For expression "PipelineExecutionStatus" we matched expected path: "Failed"
[37]:
execution.list_steps()
[37]:
[{'StepName': 'AbaloneMSEFail',
'StartTime': datetime.datetime(2022, 4, 18, 1, 11, 4, 749000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 1, 11, 5, 139000, tzinfo=tzlocal()),
'StepStatus': 'Failed',
'AttemptCount': 0,
'FailureReason': 'Execution failed due to MSE > 3.0',
'Metadata': {'Fail': {'ErrorMessage': 'Execution failed due to MSE > 3.0'}}},
{'StepName': 'AbaloneMSECond',
'StartTime': datetime.datetime(2022, 4, 18, 1, 11, 3, 792000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 1, 11, 4, 204000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'Condition': {'Outcome': 'False'}}},
{'StepName': 'AbaloneEval',
'StartTime': datetime.datetime(2022, 4, 18, 1, 6, 43, 271000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 1, 11, 2, 665000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-0xj80rygp8xr-abaloneeval-ffrpzxrakl'}}},
{'StepName': 'AbaloneTrain',
'StartTime': datetime.datetime(2022, 4, 18, 1, 3, 53, 900000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 1, 6, 42, 35000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'TrainingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:training-job/pipelines-0xj80rygp8xr-abalonetrain-u6uyl1wglo'}}},
{'StepName': 'AbaloneProcess',
'StartTime': datetime.datetime(2022, 4, 18, 0, 59, 32, 429000, tzinfo=tzlocal()),
'EndTime': datetime.datetime(2022, 4, 18, 1, 3, 53, 361000, tzinfo=tzlocal()),
'StepStatus': 'Succeeded',
'AttemptCount': 0,
'Metadata': {'ProcessingJob': {'Arn': 'arn:aws:sagemaker:us-west-2:000000000000:processing-job/pipelines-0xj80rygp8xr-abaloneprocess-cwlzrt0ero'}}}]