Summary
We started off the chapter by understanding biological and artificial neurons. Then we learned about ANNs and their layers. We learned different types of activation functions and how they are used to introduce nonlinearity in the network.
Later, we learned about the forward and backward propagation in the neural network. Next, we learned how to implement an ANN. Moving on, we learned about RNNs and how they differ from feedforward networks. Next, we learned about the variant of the RNN called LSTM. Going forward, we learned about CNNs, how they use different types of layers, and the architecture of CNNs in detail.
At the end of the chapter, we learned about an interesting algorithm called GAN. We understood the generator and discriminator component of GAN and we also explored the architecture of GAN in detail. Followed by that, we examined the loss function of GAN in detail.
In the next chapter, we will learn about one of the most popularly used deep...