When you enroll through our links, we may earn a small commission—at no extra cost to you. This helps keep our platform free and inspires us to add more value.

Deploying TensorFlow Models to AWS, Azure, and the GCP
This course will help the data scientist or engineer with a great ML model, built in TensorFlow, deploy that model to production locally or on the three major cloud platforms; Azure, AWS, or the GCP.

This Course Includes
pluralsight
3 (19 reviews )
2 hour 11 minutes
english
Online - Self Paced
core courses
pluralsight
About Deploying TensorFlow Models to AWS, Azure, and the GCP
Deploying and hosting your trained TensorFlow model locally or on your cloud platform of choice - Azure, AWS or, the GCP, can be challenging. In this course, Deploying TensorFlow Models to AWS, Azure, and the GCP, you will learn how to take your model to production on the platform of your choice. This course starts off by focusing on how you can save the model parameters of a trained model using the Saved Model interface, a universal interface for TensorFlow models. You will then learn how to scale the locally hosted model by packaging all dependencies in a Docker container. You will then get introduced to the AWS SageMaker service, the fully managed ML service offered by Amazon. Finally, you will get to work on deploying your model on the Google Cloud Platform using the Cloud ML Engine. At the end of the course, you will be familiar with how a production-ready TensorFlow model is set up as well as how to build and train your models end to end on your local machine and on the three major cloud platforms. Software required: TensorFlow, Python.
What You Will Learn?
- Course Overview : 2mins.
- Using TensorFlow Serving : 41mins.
- Containerizing TensorFlow Models Using Docker on Microsoft Azure : 26mins.
- Deploying TensorFlow Models on Amazon AWS : 25mins.
- Deploying TensorFlow Models on the Google Cloud Platform : 36mins.