Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Mastering Azure Machine Learning

You're reading from   Mastering Azure Machine Learning Perform large-scale end-to-end advanced machine learning in the cloud with Microsoft Azure Machine Learning

Arrow left icon
Product type Paperback
Published in Apr 2020
Publisher Packt
ISBN-13 9781789807554
Length 436 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Authors (2):
Arrow left icon
Christoph Körner Christoph Körner
Author Profile Icon Christoph Körner
Christoph Körner
Kaijisse Waaijer Kaijisse Waaijer
Author Profile Icon Kaijisse Waaijer
Kaijisse Waaijer
Arrow right icon
View More author details
Toc

Table of Contents (20) Chapters Close

Preface Section 1: Azure Machine Learning
1. Building an end-to-end machine learning pipeline in Azure FREE CHAPTER 2. Choosing a machine learning service in Azure Section 2: Experimentation and Data Preparation
3. Data experimentation and visualization using Azure 4. ETL, data preparation, and feature extraction 5. Azure Machine Learning pipelines 6. Advanced feature extraction with NLP Section 3: Training Machine Learning Models
7. Building ML models using Azure Machine Learning 8. Training deep neural networks on Azure 9. Hyperparameter tuning and Automated Machine Learning 10. Distributed machine learning on Azure 11. Building a recommendation engine in Azure Section 4: Optimization and Deployment of Machine Learning Models
12. Deploying and operating machine learning models 13. MLOps—DevOps for machine learning 14. What's next? Index

Summary

In this chapter, you learned how to preprocess textual and categorical nominal and ordinal data using state-of-the-art NLP techniques.

You can now build a classical NLP pipeline with stop-word removal, lemmatization and stemming, n-grams, and count term occurrences using a bag-of-words model. We used SVD to reduce the dimensionality of the resulting feature vector and to generate lower- dimensional topic encoding. One important tweak to the count-based bag-of-words model is to compare the relative term frequencies of a document. You learned about the tf-idf function and can use it to compute the importance of a word in a document compared to the corpus.

In the following section, we looked at Word2Vec and GloVe, pre-trained dictionaries of numeric word embeddings. You can now easily reuse a pre-trained word embedding for commercial NLP applications with great improvements and with accuracy due to the semantic embedding of words.

Finally, we finished the chapter by looking...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image