Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Machine Learning Engineering on AWS

You're reading from   Machine Learning Engineering on AWS Build, scale, and secure machine learning systems and MLOps pipelines in production

Arrow left icon
Product type Paperback
Published in Oct 2022
Publisher Packt
ISBN-13 9781803247595
Length 530 pages
Edition 1st Edition
Tools
Arrow right icon
Author (1):
Arrow left icon
Joshua Arvin Lat Joshua Arvin Lat
Author Profile Icon Joshua Arvin Lat
Joshua Arvin Lat
Arrow right icon
View More author details
Toc

Table of Contents (19) Chapters Close

Preface 1. Part 1: Getting Started with Machine Learning Engineering on AWS
2. Chapter 1: Introduction to ML Engineering on AWS FREE CHAPTER 3. Chapter 2: Deep Learning AMIs 4. Chapter 3: Deep Learning Containers 5. Part 2:Solving Data Engineering and Analysis Requirements
6. Chapter 4: Serverless Data Management on AWS 7. Chapter 5: Pragmatic Data Processing and Analysis 8. Part 3: Diving Deeper with Relevant Model Training and Deployment Solutions
9. Chapter 6: SageMaker Training and Debugging Solutions 10. Chapter 7: SageMaker Deployment Solutions 11. Part 4:Securing, Monitoring, and Managing Machine Learning Systems and Environments
12. Chapter 8: Model Monitoring and Management Solutions 13. Chapter 9: Security, Governance, and Compliance Strategies 14. Part 5:Designing and Building End-to-end MLOps Pipelines
15. Chapter 10: Machine Learning Pipelines with Kubeflow on Amazon EKS 16. Chapter 11: Machine Learning Pipelines with SageMaker Pipelines 17. Index 18. Other Books You May Enjoy

Using the Kubeflow Pipelines SDK to build ML workflows

In this section, we will build ML workflows using the Kubeflow Pipelines SDK. The Kubeflow Pipelines SDK contains what we need to build the pipeline components containing the custom code we want to run. Using the Kubeflow Pipelines SDK, we can define the Python functions that would map to the pipeline components of a pipeline.

Here are some guidelines that we need to follow when building Python function-based components using the Kubeflow Pipelines SDK:

  • The defined Python functions should be standalone and should not use any code and variables declared outside of the function definition. This means that import statements (for example, import pandas) should be implemented inside the function, too. Here’s a quick example of how imports should be implemented:
    def process_data(...):
        import pandas as pd    
        df_all_data = pd.read_csv(df_all_data_path...
lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime