Implementing end-to-end language models
In the previous sections, we trained and concatenated multiple pieces to implement a final algorithm where most of the individual steps need to be trained as well. Lemmatization contains a dictionary of conversion rules. Stop words are stored in the dictionary. Stemming needs rules for each language and word that the embedding needs to train—TF-IDF and SVD are only computed on your training data but are independent of each other.
This is a similar problem to the traditional computer vision approach, which we will discuss in more depth in Chapter 10, Training Deep Neural Networks on Azure, where many classic algorithms are combined into a pipeline of feature extractors and classifiers. Similar to breakthroughs of end-to-end models trained via gradient descent and backpropagation in computer vision, deep neural networks—especially sequence-to-sequence models—have replaced the classical approach of performing each step of...