Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Python: Advanced Guide to Artificial Intelligence

You're reading from   Python: Advanced Guide to Artificial Intelligence Expert machine learning systems and intelligent agents using Python

Arrow left icon
Product type Course
Published in Dec 2018
Publisher Packt
ISBN-13 9781789957211
Length 764 pages
Edition 1st Edition
Languages
Arrow right icon
Authors (2):
Arrow left icon
Giuseppe Bonaccorso Giuseppe Bonaccorso
Author Profile Icon Giuseppe Bonaccorso
Giuseppe Bonaccorso
Rajalingappaa Shanmugamani Rajalingappaa Shanmugamani
Author Profile Icon Rajalingappaa Shanmugamani
Rajalingappaa Shanmugamani
Arrow right icon
View More author details
Toc

Table of Contents (31) Chapters Close

Title Page
About Packt
Contributors
Preface
1. Machine Learning Model Fundamentals FREE CHAPTER 2. Introduction to Semi-Supervised Learning 3. Graph-Based Semi-Supervised Learning 4. Bayesian Networks and Hidden Markov Models 5. EM Algorithm and Applications 6. Hebbian Learning and Self-Organizing Maps 7. Clustering Algorithms 8. Advanced Neural Models 9. Classical Machine Learning with TensorFlow 10. Neural Networks and MLP with TensorFlow and Keras 11. RNN with TensorFlow and Keras 12. CNN with TensorFlow and Keras 13. Autoencoder with TensorFlow and Keras 14. TensorFlow Models in Production with TF Serving 15. Deep Reinforcement Learning 16. Generative Adversarial Networks 17. Distributed Models with TensorFlow Clusters 18. Debugging TensorFlow Models 19. Tensor Processing Units
20. Getting Started 21. Image Classification 22. Image Retrieval 23. Object Detection 24. Semantic Segmentation 25. Similarity Learning 1. Other Books You May Enjoy Index

Simple Recurrent Neural Network


Here is what a simple neural network with loops looks like:

RNN Network

In this diagram, a Neural Network N takes input

 to produce output

. Due to the loop, at the next time step

, it takes the input

 along with input

 to produce output

. Mathematically, we represent this as the following equation:

When we unroll the loop, the RNN architecture looks as follows at time step

:

Unrolled RNN at timestep t1

As the time steps evolve, this loop unrolls as follows at time step 5:

Unrolled RNN at timestep t5

At every time step, the same learning function,

, and the same parameters, w and b, are used.

The output y is not always produced at every time step. Instead, an output h is produced at every time step, and another activation function is applied to this output h to produce the output y. The equations for the RNN look like this now:

where,

  • is the weight vector for x inputs that are connected to the hidden layer
  • is the weight vector for the value of h from the previous time...
lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime
Banner background image