MLOps focuses on realizing the value from any data science activity through the adoption of DevOps and engineering best practices to build solutions and products that have an ML component at their core. It provides the standards, tools, and framework that supports the work of data science teams to take their ideas from the whiteboard to production in a timely, secure, traceable, and repeatable manner.
With this MLOps architecture, AWS partners can enable their customers to accelerate operationalizing ML models in production to achieve the benefits of an automated machine learning lifecycle management by reducing the typical lifecycle for a significantly faster delivery time.
The solutions architecture here was made to address a particular use-case (people analytics attrition model) but the intention was to design it in such a way that it can be used generically across different ML problems. Of course, this will not meet all ML use-case requirements but many of the components are key to almost all ML systems in AWS (S3 storage, SageMaker hyperparameter tuning, training, and deployment). This is intended to be a starting point for establishing proper ML pipeline automation with many re-usable elements that can form the foundation for a mature ML system.
This architecture describes using machine learning (ML) for the automated identification of unhappy customers, also known as customer churn prediction. Losing customers is costly for any business, so identifying unhappy customers early on gives you a chance to offer them incentives to stay.
We shall use a familiar example of churn: leaving a mobile phone operator. And if the provider knows that a customer is thinking of leaving, it can offer timely incentives - such as a phone upgrade or perhaps having a new feature activated – and the customer may stick around. Incentives are often much more cost-effective than losing and reacquiring a customer.
The architecture shown above is an example of an automated MLOps pipeline could look like in AWS. This is for a batch inference model deployment where a persistent endpoint is not required (although it is optionally available as part of the infrastructure build). It uses a serverless architecture which has benefits in terms of cost-efficiency and ease of development.
The key AWS services are:
- AWS CodePipeline automates your software release process, allowing you to rapidly release new features to your users. With CodePipeline, you can quickly iterate on feedback and get new features to your users faster.
- AWS CodeBuild eliminates the need to set up, patch, update, and manage your own build servers and software. There is no software to install or manage.
- Amazon S3 for initial data storage/data lakes, storing flat-file data extracts, source code, model objects, inference output, and metadata storage.
- Amazon SageMaker Amazon SageMaker helps you streamline the machine learning (ML) lifecycle by automating and standardizing MLOps practices across your organization. You can easily build, train, deploy, and manage ML models, whether it’s only a few, hundreds of thousands, or even millions. With purpose-built tools for ML lifecycle management and built-in integrations with other AWS services, you can boost productivity of data scientists and ML engineers while maintaining high model accuracy and enhancing security and compliance.
- Amazon CloudWatch for monitoring SageMaker tuning and training jobs.
We have received a lot of feedback from a number of AWS Strategic partners and customers about their MLOps use cases and discovered that they are using Terraform for their infrastructure automation. As a result of this feedback, we decided to use Terraform to help them in their efforts. Here are additional reasons for using Terraform for MLOps use cases:
- Terraform can define infrastructure and manage infrastructures’ lifecycle directly. To use SageMaker Studio project, a project template should be created in advance and saved in Service Catalog. The template needs to define infrastructure using CloudFormation.
- Terraform supports multi-cloud platforms. User can provision services on multiple cloud platform.
- Terraform has workspaces, which makes it easier to manage multiple environments.
- Using Terraform can expedite deployment process because development team lacks skills and experiences in CloudFormation.
- Terraform CLI installed on your local machine or cloud workspace (if you use cloud workspace)
- AWS CLI installed
- An AWS account
- Your AWS credentials.
- Github repositories. One repository is for build repository (“modelbuild-pipeline”), and the other one is for deploy repository (“modeldeploy-pipeline”). The repository names match those in the variable file.
Terraform Cloud is an application that manages Terraform runs in a consistent and reliable environment instead of on your local machine. It stores shared state and secret data, and connects to version control systems so that you and your team can work on infrastructure as code within your usual code workflow. It also has a private registry for sharing Terraform modules.
Visit https://app.terraform.io/signup/account and follow the prompts to create a free Terraform Cloud account.
When you sign up, you'll receive an email asking you to confirm your email address. Confirm your email address before moving on.
After you create your account, the Terraform Cloud Web UI will prompt you to create a new organization. Your organization is free, and the members you add will be able to collaborate on your workspaces and share private modules.
Enter your own unique organization name
(here 'aws-terraform'
as example) and an email address. You can use the same email address that you used for your account.
Workspaces determine how Terraform Cloud organizes infrastructure.
Click CLI-driven workflow.
Enter your own unique Workspace name
(here aws-terraform-lambda-container
as example) and click Create workspace.
Copy the Example code and paste it inside the main.tf file. The organization and workspace names are those you created in the previous steps.
terraform {
cloud {
organization = "<your organization name>"
workspaces {
name = "<your workspace name>"
}
}
}
Click the Variables tab at the top to create Terraform Variables AWS_REGION and Environment Variables AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY.
git clone https://github.com/aws-samples/aws-mlops-pipelines-terraform.git
There is 'Notebooks' folder, which contains two jupyter notebook examples. One uses SageMaker workflow pipeline within the notebook. The other one uses SageMaker SDK without pipeline.
Push “modelbuild-pipeline” folder contents to your own Github repository “modelbuild-pipeline”.
Push “modeldeploy-pipeline” folder to your own Github repository “modeldeploy-pipeline”.
In your workspace or local machine, open a new Terminal,go to the 'terraform' folder, and use the following command:
cd ~/terraform
terraform login
At the prompt, enter 'yes' and follow the instructions to generate a token. Terraform will store the token in your workspace or local machine.
Go to your terraform code folder “terraform” run the following commands.
terraform init
terraform plan
terraform apply
Go to your AWS CodePipeline console, you can check the pipeline running progress.
To invoke and test the endpoint, you can copy 'test.csv' from 'Notebooks' folder to your own notebook folder, and run the following code within a notebook.
import pandas as pd
import numpy as np
import sagemaker
import boto3
from sagemaker import get_execution_role
test_data=pd.read_csv('test.csv',header=None)
testdata1=test_data.iloc[0:1,1:]
runtime = boto3.client("sagemaker-runtime")
Endpoint_name=<your endpoint name> # update to your own endpoint name
prediction = runtime.invoke_endpoint(
EndpointName=Endpoint_name,
Body=testdata1.to_csv(header=False, index=False).encode("utf-8"),
ContentType="text/csv",
Accept= "text/csv",
)
print(prediction["Body"].read())
To avoid incurring future charges, please execute the following clean-up steps. Login to the AWS Console and enter your credentials
- Select Services from the top menu, and choose Amazon SageMaker Endpoints Go to Inference / Endpoints on the left-hand menu, click the radio button next to each endpoint. Select Actions and then Delete, then confirm delete hitting the Delete button.
- Select Services from the top menu, and choose S3 Delete all S3 buckets created for this project
- Remove AWS infrastructure, run the following command from your “terraform” folder
terraform destroy
At the prompt, enter 'yes' and follow the instructions to destroy resources.