Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Save more on your purchases! discount-offer-chevron-icon
Savings automatically calculated. No voucher code required.
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Practical Deep Learning at Scale with MLflow

You're reading from   Practical Deep Learning at Scale with MLflow Bridge the gap between offline experimentation and online production

Arrow left icon
Product type Paperback
Published in Jul 2022
Publisher Packt
ISBN-13 9781803241333
Length 288 pages
Edition 1st Edition
Tools
Arrow right icon
Author (1):
Arrow left icon
Yong Liu Yong Liu
Author Profile Icon Yong Liu
Yong Liu
Arrow right icon
View More author details
Toc

Table of Contents (17) Chapters Close

Preface 1. Section 1 - Deep Learning Challenges and MLflow Prime
2. Chapter 1: Deep Learning Life Cycle and MLOps Challenges FREE CHAPTER 3. Chapter 2: Getting Started with MLflow for Deep Learning 4. Section 2 –
Tracking a Deep Learning Pipeline at Scale
5. Chapter 3: Tracking Models, Parameters, and Metrics 6. Chapter 4: Tracking Code and Data Versioning 7. Section 3 –
Running Deep Learning Pipelines at Scale
8. Chapter 5: Running DL Pipelines in Different Environments 9. Chapter 6: Running Hyperparameter Tuning at Scale 10. Section 4 –
Deploying a Deep Learning Pipeline at Scale
11. Chapter 7: Multi-Step Deep Learning Inference Pipeline 12. Chapter 8: Deploying a DL Inference Pipeline at Scale 13. Section 5 – Deep Learning Model Explainability at Scale
14. Chapter 9: Fundamentals of Deep Learning Explainability 15. Chapter 10: Implementing DL Explainability with MLflow 16. Other Books You May Enjoy

Summary

In this chapter, we have learned different ways to deploy an MLflow inference pipeline model for both batch inference and online real-time inference. We started with a brief survey on different model serving scenarios (batch, streaming, and on-device) and looked at three different categories of tools for MLflow model deployment (the MLflow built-in deployment tool, MLflow deployment plugins, and generic model inference serving frameworks that could work with the MLflow inference model). Then, we covered several local deployment scenarios, using the PySpark UDF function to do batch inference and MLflow local deployment for web service. Afterward, we learned how to use Ray Serve in conjunction with the mlflow-ray-serve plugin to deploy an MLflow Python inference pipeline model into a local Ray cluster. This opens doors to deploy to any cloud platform such as AWS, Azure ML, or GCP, as long as we can set up a Ray cluster in the cloud. Finally, we provided a complete end-to-end...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image