Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Codeless Deep Learning with KNIME

You're reading from   Codeless Deep Learning with KNIME Build, train, and deploy various deep neural network architectures using KNIME Analytics Platform

Arrow left icon
Product type Paperback
Published in Nov 2020
Publisher Packt
ISBN-13 9781800566613
Length 384 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Authors (3):
Arrow left icon
Kathrin Melcher Kathrin Melcher
Author Profile Icon Kathrin Melcher
Kathrin Melcher
KNIME AG KNIME AG
Author Profile Icon KNIME AG
KNIME AG
Rosaria Silipo Rosaria Silipo
Author Profile Icon Rosaria Silipo
Rosaria Silipo
Arrow right icon
View More author details
Toc

Table of Contents (16) Chapters Close

Preface 1. Section 1: Feedforward Neural Networks and KNIME Deep Learning Extension
2. Chapter 1: Introduction to Deep Learning with KNIME Analytics Platform FREE CHAPTER 3. Chapter 2: Data Access and Preprocessing with KNIME Analytics Platform 4. Chapter 3: Getting Started with Neural Networks 5. Chapter 4: Building and Training a Feedforward Neural Network 6. Section 2: Deep Learning Networks
7. Chapter 5: Autoencoder for Fraud Detection 8. Chapter 6: Recurrent Neural Networks for Demand Prediction 9. Chapter 7: Implementing NLP Applications 10. Chapter 8: Neural Machine Translation 11. Chapter 9: Convolutional Neural Networks for Image Classification 12. Section 3: Deployment and Productionizing
13. Chapter 10: Deploying a Deep Learning Network 14. Chapter 11: Best Practices and Other Deployment Options 15. Other Books You May Enjoy

Chapter 6: Recurrent Neural Networks for Demand Prediction

We have gathered some experience, by now, with fully connected feedforward neural networks in two variants: implementing a classification task by assigning an input sample to a class in a set of predefined classes or trying to reproduce the shape of an input vector via an autoencoder architecture. In both cases, the output response depends only on the values of the current input vector. At time , the output response, , depends on, and only on, the input vector, , at time . The network has no memory of what came before and produces only based on input .

With Recurrent Neural Networks (RNNs), we introduce the time component . We are going to discover networks where the output response, , at time depends on the current input sample, , as well as on previous input samples, , , … , where the memory of the network of the past samples depends on the network architecture.

We will first introduce the general concept...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime