Search icon CANCEL
Subscription
0
Cart icon
Cart
Close icon
You have no products in your basket yet
Save more on your purchases!
Savings automatically calculated. No voucher code required
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Ensemble Machine Learning Cookbook

You're reading from  Ensemble Machine Learning Cookbook

Product type Book
Published in Jan 2019
Publisher Packt
ISBN-13 9781789136609
Pages 336 pages
Edition 1st Edition
Languages
Authors (2):
Dipayan Sarkar Dipayan Sarkar
Profile icon Dipayan Sarkar
Vijayalakshmi Natarajan Vijayalakshmi Natarajan
Profile icon Vijayalakshmi Natarajan
View More author details
Toc

Table of Contents (14) Chapters close

Preface 1. Get Closer to Your Data 2. Getting Started with Ensemble Machine Learning 3. Resampling Methods 4. Statistical and Machine Learning Algorithms 5. Bag the Models with Bagging 6. When in Doubt, Use Random Forests 7. Boosting Model Performance with Boosting 8. Blend It with Stacking 9. Homogeneous Ensembles Using Keras 10. Heterogeneous Ensemble Classifiers Using H2O 11. Heterogeneous Ensemble for Text Classification Using NLP 12. Homogenous Ensemble for Multiclass Classification Using Keras 13. Other Books You May Enjoy

What this book covers

Chapter 1, Get Closer to Your Data, explores a dataset and implements hands-on coding with Python for exploratory data analysis using statistical methods and visualization for the dataset.

Chapter 2, Getting Started with Ensemble Machine Learning, explores what ensemble learning is and how it can help in real-life scenarios. Basic ensemble techniques, including averaging, weighted averaging, and max-voting, are explained. These techniques form the basis for ensemble techniques, and an understanding of them will lay the groundwork for readers to move to more advanced stage after reading this chapter.

Chapter 3, Resampling Methods, introduces a handful of algorithms that will be useful when we get into an ensemble of multiple heterogeneous algorithms. This chapter uses scikit-learn to prepare all the algorithms to be used.

Chapter 4, Statistical and Machine Learning Algorithms, helps the readers to get to know various types of resampling methods that are used by machine-learning algorithms. Each resampling method has its advantages and disadvantages, which are explained to the readers. The readers also learn the code to be executed for each type of sampling.

Chapter 5, Bag the Models with Bagging, provides the readers with an understanding of what bootstrap aggregation is and how the bootstrap results can be aggregated, in a process also known as bagging.

Chapter 6, When in Doubt, Use Random Forests, introduces the random forest algorithm. It will introduce to readers how, and what kind of, ensemble techniques are used by Random Forest and how this helps our models avoid overfitting.

Chapter 7, Boosting Model Performance with Boosting, introduces boosting and discusses how it helps to improve a model performance by reducing variances and increasing accuracy. This chapter provides information such as the fact that boosting is not robust against outliers and noisy data but is flexible and can be used with a loss function.

Chapter 8, Blend It with Stacking, applies stacking to learn the optimal combination of base learners. This chapter will acquaint readers with stacking, which is also known as stacked generalization.

Chapter 9, Homogeneous Ensemble Using Keras, is a complete code walk-through on a classification case study for recognizing hand-written digits with homogeneous algorithms – in this case, multiple neural network models using Keras.

Chapter 10, Heterogeneous Ensemble Classifiers Using H2O, is a complete code walk-through on a classification case study for default prediction with an ensemble of multiple heterogeneous algorithms using scikit-learn.

Chapter 11, Heterogeneous Ensemble for Text Classification Using NLP, is a complete code walk-through on a classification case study to classify sentiment polarity using an ensemble of multiple heterogeneous algorithms. Here, NLP techniques such as semantics are used to improve the accuracy of classification. Then, the mined text information is used to employ ensemble classification techniques for sentiment analysis. In this case study, the H2O library is used for building models.

Chapter 12, Homogeneous Ensemble for Multiclass Classification Using Keras, is a complete code walk-through on a classification case study for multiple classification with homogeneous ensemble using data diversity with the tf.keras module from TensorFlow.

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $15.99/month. Cancel anytime}