Search icon CANCEL
Subscription
0
Cart icon
Close icon
You have no products in your basket yet
Save more on your purchases!
Savings automatically calculated. No voucher code required
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Responsible AI in the Enterprise

You're reading from  Responsible AI in the Enterprise

Product type Book
Published in Jul 2023
Publisher Packt
ISBN-13 9781803230528
Pages 318 pages
Edition 1st Edition
Languages
Authors (2):
Adnan Masood Adnan Masood
Profile icon Adnan Masood
Heather Dawe Heather Dawe
Profile icon Heather Dawe
View More author details

Table of Contents (16) Chapters

Preface 1. Part 1: Bigot in the Machine – A Primer
2. Chapter 1: Explainable and Ethical AI Primer 3. Chapter 2: Algorithms Gone Wild 4. Part 2: Enterprise Risk Observability Model Governance
5. Chapter 3: Opening the Algorithmic Black Box 6. Chapter 4: Robust ML – Monitoring and Management 7. Chapter 5: Model Governance, Audit, and Compliance 8. Chapter 6: Enterprise Starter Kit for Fairness, Accountability, and Transparency 9. Part 3: Explainable AI in Action
10. Chapter 7: Interpretability Toolkits and Fairness Measures – AWS, GCP, Azure, and AIF 360 11. Chapter 8: Fairness in AI Systems with Microsoft Fairlearn 12. Chapter 9: Fairness Assessment and Bias Mitigation with Fairlearn and the Responsible AI Toolbox 13. Chapter 10: Foundational Models and Azure OpenAI 14. Index 15. Other Books You May Enjoy

Getting started with fairness

The Fairlearn toolkit is an open source tool for assessing and improving the fairness of AI systems built by data scientists and developers. Fairlearn includes a visualization dashboard and algorithms for mitigating unfairness, along with required metrics. As AI and ML algorithms increasingly shape our world, it is critical that we ensure fairness in their application by using tools that can identify and mitigate bias. Fairlearn is one such library. As we dive into the use of Fairlearn, we must understand the reasons why it is important to consider the potential impact of sensitive features on your ML models, even if you are not explicitly including sensitive features in the training data.

A common misconception is “If we remove sensitive features such as a person’s race, sex, religion, sexual orientation, veteran status, and so on, shouldn’t that be enough to mitigate any bias?” The answer is “Not really” because...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at ₹800/month. Cancel anytime}