Search icon CANCEL
Subscription
0
Cart icon
Cart
Close icon
You have no products in your basket yet
Save more on your purchases!
Savings automatically calculated. No voucher code required
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Responsible AI in the Enterprise

You're reading from  Responsible AI in the Enterprise

Product type Book
Published in Jul 2023
Publisher Packt
ISBN-13 9781803230528
Pages 318 pages
Edition 1st Edition
Languages
Authors (2):
Adnan Masood Adnan Masood
Profile icon Adnan Masood
Heather Dawe Heather Dawe
Profile icon Heather Dawe
View More author details
Toc

Table of Contents (16) Chapters close

Preface 1. Part 1: Bigot in the Machine – A Primer
2. Chapter 1: Explainable and Ethical AI Primer 3. Chapter 2: Algorithms Gone Wild 4. Part 2: Enterprise Risk Observability Model Governance
5. Chapter 3: Opening the Algorithmic Black Box 6. Chapter 4: Robust ML – Monitoring and Management 7. Chapter 5: Model Governance, Audit, and Compliance 8. Chapter 6: Enterprise Starter Kit for Fairness, Accountability, and Transparency 9. Part 3: Explainable AI in Action
10. Chapter 7: Interpretability Toolkits and Fairness Measures – AWS, GCP, Azure, and AIF 360 11. Chapter 8: Fairness in AI Systems with Microsoft Fairlearn 12. Chapter 9: Fairness Assessment and Bias Mitigation with Fairlearn and the Responsible AI Toolbox 13. Chapter 10: Foundational Models and Azure OpenAI 14. Index 15. Other Books You May Enjoy

Summary

Data and model drift refer to a phenomenon that occurs when the statistical properties of a dataset or underlying model change over time. In this chapter, we reviewed how this can have an adverse impact on the predictions of models and, hence, on business outcomes. To make sure models function as desired, companies implement an ML life cycle that ensures design, development, deployment, and monitoring best practices are in place. Drifts can happen for a variety of reasons, including changes in the underlying population and changes in the way data is collected. When data drift happens, it can create bias in ML models that are trained on this data, which can be quite problematic for regulations and compliance.

In this chapter, we reviewed several ways to detect and mitigate bias due to data or model drift, and to monitor your training and validation error rates closely using different tools, including open source and commercial hyperscaler products. There are various other...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $15.99/month. Cancel anytime