Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Codeless Deep Learning with KNIME

You're reading from   Codeless Deep Learning with KNIME Build, train, and deploy various deep neural network architectures using KNIME Analytics Platform

Arrow left icon
Product type Paperback
Published in Nov 2020
Publisher Packt
ISBN-13 9781800566613
Length 384 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Authors (3):
Arrow left icon
Kathrin Melcher Kathrin Melcher
Author Profile Icon Kathrin Melcher
Kathrin Melcher
KNIME AG KNIME AG
Author Profile Icon KNIME AG
KNIME AG
Rosaria Silipo Rosaria Silipo
Author Profile Icon Rosaria Silipo
Rosaria Silipo
Arrow right icon
View More author details
Toc

Table of Contents (16) Chapters Close

Preface 1. Section 1: Feedforward Neural Networks and KNIME Deep Learning Extension
2. Chapter 1: Introduction to Deep Learning with KNIME Analytics Platform FREE CHAPTER 3. Chapter 2: Data Access and Preprocessing with KNIME Analytics Platform 4. Chapter 3: Getting Started with Neural Networks 5. Chapter 4: Building and Training a Feedforward Neural Network 6. Section 2: Deep Learning Networks
7. Chapter 5: Autoencoder for Fraud Detection 8. Chapter 6: Recurrent Neural Networks for Demand Prediction 9. Chapter 7: Implementing NLP Applications 10. Chapter 8: Neural Machine Translation 11. Chapter 9: Convolutional Neural Networks for Image Classification 12. Section 3: Deployment and Productionizing
13. Chapter 10: Deploying a Deep Learning Network 14. Chapter 11: Best Practices and Other Deployment Options 15. Other Books You May Enjoy

Introduction to transfer learning

The general idea of transfer learning is to reuse the knowledge gained by a network trained for task A on another related task B. For example, if we train a network to recognize sailing boats (task A), we can use this network as a starting point to train a new model to recognize motorboats (task B). In this case, task A is called the source task and task B the target task.

Reusing a trained network as the starting point to train a new network is different from the traditional way of training networks, whereby neural networks are trained on their own for specific tasks on specific datasets. Figure 9.19 here visualizes the traditional way of network training, whereby different systems are trained for different tasks and domains:

Figure 9.19 – Traditional way of training machine learning models and neural networks

Figure 9.19 – Traditional way of training machine learning models and neural networks

But why should we use transfer learning instead of training models in the traditional, isolated way...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime