Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Hands-On Data Analysis with Pandas

You're reading from   Hands-On Data Analysis with Pandas A Python data science handbook for data collection, wrangling, analysis, and visualization

Arrow left icon
Product type Paperback
Published in Apr 2021
Publisher Packt
ISBN-13 9781800563452
Length 788 pages
Edition 2nd Edition
Languages
Tools
Arrow right icon
Author (1):
Arrow left icon
Stefanie Molin Stefanie Molin
Author Profile Icon Stefanie Molin
Stefanie Molin
Arrow right icon
View More author details
Toc

Table of Contents (21) Chapters Close

Preface 1. Section 1: Getting Started with Pandas
2. Chapter 1: Introduction to Data Analysis FREE CHAPTER 3. Chapter 2: Working with Pandas DataFrames 4. Section 2: Using Pandas for Data Analysis
5. Chapter 3: Data Wrangling with Pandas 6. Chapter 4: Aggregating Pandas DataFrames 7. Chapter 5: Visualizing Data with Pandas and Matplotlib 8. Chapter 6: Plotting with Seaborn and Customization Techniques 9. Section 3: Applications – Real-World Analyses Using Pandas
10. Chapter 7: Financial Analysis – Bitcoin and the Stock Market 11. Chapter 8: Rule-Based Anomaly Detection 12. Section 4: Introduction to Machine Learning with Scikit-Learn
13. Chapter 9: Getting Started with Machine Learning in Python 14. Chapter 10: Making Better Predictions – Optimizing Models 15. Chapter 11: Machine Learning Anomaly Detection 16. Section 5: Additional Resources
17. Chapter 12: The Road Ahead 18. Solutions
19. Other Books You May Enjoy Appendix

Ensemble methods

Ensemble methods combine many models (often weak ones) to create a stronger one that will either minimize the average error between observed and predicted values (the bias) or improve how well it generalizes to unseen data (minimize the variance). We have to strike a balance between complex models that may increase variance, as they tend to overfit, and simple models that may have high bias, as these tend to underfit. This is called the bias-variance trade-off, which is illustrated in the following subplots:

Figure 10.11 – The bias-variance trade-off

Ensemble methods can be broken down into three categories: boosting, bagging, and stacking. Boosting trains many weak learners, which learn from each other's mistakes to reduce bias, making a stronger learner. Bagging, on the other hand, uses bootstrap aggregation to train many models on bootstrap samples of the data and aggregate the results together (using voting for classification...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime