TensorFlow™ enables developers to quickly and easily get started with deep learning in the cloud. The framework has broad support in the industry and has become a popular choice for deep learning research and application development, particularly in areas such as computer vision, natural language understanding and speech translation. You can get started on AWS with a fully-managed TensorFlow experience with Amazon SageMaker, a platform to build, train, and deploy machine learning models at scale.
The Amazon SageMaker Python SDK provides open source APIs and containers that make it easy to train and deploy models in Amazon SageMaker with several different machine learning and deep learning frameworks. For general information about the Amazon SageMaker Python SDK, see https://sagemaker.readthedocs.io/.
You can use Amazon SageMaker to train and deploy a model using custom TensorFlow code. The Amazon SageMaker Python SDK TensorFlow estimators and models and the Amazon SageMaker open-source TensorFlow containers make writing a TensorFlow script and running it in Amazon SageMaker easier.
In this workshop you will port a working TensorFlow script to run on SageMaker and utilize some of the feature available for TensorFlow in SageMaker
The workshop is based on 5 modules:
- Porting a TensorFlow script to run in SageMaker using SageMaker script mode.
- Monitoring your training job using TensorBoard and Amazon CloudWatch metrics.
- Optimizing your training job using SageMaker pipemode input.
- Running a distributed training job.
- Deploying your trained model on Amazon SageMaker.
This sample code is made available under the MIT-0 license. See the LICENSE file.