Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
R Machine Learning Projects

You're reading from   R Machine Learning Projects Implement supervised, unsupervised, and reinforcement learning techniques using R 3.5

Arrow left icon
Product type Paperback
Published in Jan 2019
Publisher Packt
ISBN-13 9781789807943
Length 334 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Author (1):
Arrow left icon
Dr. Sunil Kumar Chinnamgari Dr. Sunil Kumar Chinnamgari
Author Profile Icon Dr. Sunil Kumar Chinnamgari
Dr. Sunil Kumar Chinnamgari
Arrow right icon
View More author details
Toc

Table of Contents (12) Chapters Close

Preface 1. Exploring the Machine Learning Landscape FREE CHAPTER 2. Predicting Employee Attrition Using Ensemble Models 3. Implementing a Jokes Recommendation Engine 4. Sentiment Analysis of Amazon Reviews with NLP 5. Customer Segmentation Using Wholesale Data 6. Image Recognition Using Deep Neural Networks 7. Credit Card Fraud Detection Using Autoencoders 8. Automatic Prose Generation with Recurrent Neural Networks 9. Winning the Casino Slot Machines with Reinforcement Learning 10. The Road Ahead
11. Other Books You May Enjoy

Autoencoders explained

Autoencoders (AEs) are neural networks that are of a feedforward and non-recurrent type. They aim to copy the given inputs to the outputs. An AE works by compressing the input into a lower dimensional summary. This summary is often referred as latent space representation. An AE attempts to reconstruct the output from the latent space representation. An Encoder, a Latent Space Representation, and a Decoder are the three parts that make up the AEs. The following figure is an illustration showing the application of an AE on a sample picked from the MNIST dataset:

Application of AE on MNIST dataset sample

The encoder and decoder components of AEs are fully-connected feedforward networks. The number of neurons in a latent space representation is a hyperparameter that needs to be passed as part of building the AE. The number of neurons or nodes that is decided...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at €18.99/month. Cancel anytime