Summary
We have covered a wide range of feature engineering techniques in this chapter. We used tools to drop redundant or highly correlated features. We explored the most common kinds of encoding—one-hot, ordinal, and hashing encoding. We then used transformations to improve the distribution of our features. Finally, we used common binning and scaling approaches to address skew, kurtosis, and outliers, and to adjust for features with widely different ranges. In the next chapter, we’ll learn how to fix messy data when aggregating.
Leave a review!
Enjoying this book? Help readers like you by leaving an Amazon review. Scan the QR code below to get a free eBook of your choice.