Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Learn TensorFlow Enterprise

You're reading from   Learn TensorFlow Enterprise Build, manage, and scale machine learning workloads seamlessly using Google's TensorFlow Enterprise

Arrow left icon
Product type Paperback
Published in Nov 2020
Publisher Packt
ISBN-13 9781800209145
Length 314 pages
Edition 1st Edition
Languages
Arrow right icon
Author (1):
Arrow left icon
KC Tung KC Tung
Author Profile Icon KC Tung
KC Tung
Arrow right icon
View More author details
Toc

Table of Contents (15) Chapters Close

Preface 1. Section 1 – TensorFlow Enterprise Services and Features
2. Chapter 1: Overview of TensorFlow Enterprise FREE CHAPTER 3. Chapter 2: Running TensorFlow Enterprise in Google AI Platform 4. Section 2 – Data Preprocessing and Modeling
5. Chapter 3: Data Preparation and Manipulation Techniques 6. Chapter 4: Reusable Models and Scalable Data Pipelines 7. Section 3 – Scaling and Tuning ML Works
8. Chapter 5: Training at Scale 9. Chapter 6: Hyperparameter Tuning 10. Section 4 – Model Optimization and Deployment
11. Chapter 7: Model Optimization 12. Chapter 8: Best Practices for Model Training and Performance 13. Chapter 9: Serving a TensorFlow Model 14. Other Books You May Enjoy

Submitting tuning jobs in Google's AI Platform

Now we are ready to use Google's AI Platform to perform hyperparameter training. You may download everything you need from the GitHub repository for this chapter. For the AI Platform code in this section, you can refer to the gcptuningwork file in this chapter's folder in the GitHub repository for the book.

In the cloud, we have access to powerful machines that can speed up our search process. Overall, the approach we will leverage is very similar to what we saw in the previous section about submitting a local Python script training job. We will use the tf.compat.v1.flag method to handle user input or flags. The rest of the script follows a similar structure, with the exception of data handling, because we will use TFRecord instead of ImageGenerator and a conditional flag for the distributed training strategy.

Since the tuning job is submitted to AI Platform from a remote node (that is, your local compute environment...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image