Search icon CANCEL
Subscription
0
Cart icon
Cart
Close icon
You have no products in your basket yet
Save more on your purchases!
Savings automatically calculated. No voucher code required
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Machine Learning Algorithms - Second Edition

You're reading from  Machine Learning Algorithms - Second Edition

Product type Book
Published in Aug 2018
Publisher Packt
ISBN-13 9781789347999
Pages 522 pages
Edition 2nd Edition
Languages
Toc

Table of Contents (19) Chapters close

Preface 1. A Gentle Introduction to Machine Learning 2. Important Elements in Machine Learning 3. Feature Selection and Feature Engineering 4. Regression Algorithms 5. Linear Classification Algorithms 6. Naive Bayes and Discriminant Analysis 7. Support Vector Machines 8. Decision Trees and Ensemble Learning 9. Clustering Fundamentals 10. Advanced Clustering 11. Hierarchical Clustering 12. Introducing Recommendation Systems 13. Introducing Natural Language Processing 14. Topic Modeling and Sentiment Analysis in NLP 15. Introducing Neural Networks 16. Advanced Deep Learning Models 17. Creating a Machine Learning Architecture 18. Other Books You May Enjoy

Summary

In this chapter, we introduced the important concepts of linear models and described how linear regression works. In particular, we focused on the basic model and its main variants: Lasso, Ridge, and ElasticNet. They don't modify the internal dynamics, but work as normalizers for the weights to avoid common problems when the dataset contains unscaled samples. These penalties have specific peculiarities. While Lasso promotes sparsity, Ridge tries to find a minimum with the constraint that the weights must lie within a circle centered at the origin (whose radius is parametrized to increase/decrease the normalization strength). ElasticNet is a mix of both these techniques, and it tries to find a minimum where the weights are small enough and a certain degree of sparsity is achieved.

We also discussed advanced techniques such as RANSAC, which allows us to cope with outliers...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $15.99/month. Cancel anytime