Audio and music
We have used CNNs for images, videos, and text. Now let's have a look to how variants of CNNs can be used for audio.
So, you might wonder why learning to synthesize audio is so difficult. Well, each digital sound we hear is based on 16,000 samples per second (sometimes 48,000 or more) and building a predictive model where we learn to reproduce a sample based on all the previous ones is a very difficult challenge.
Dilated ConvNets, WaveNet, and NSynth
WaveNet is a deep generative model for producing raw audio waveforms. This breakthrough technology has been introduced (WaveNet is available at https://deepmind.com/blog/wavenet-generative-model-raw-audio/) by Google DeepMind for teaching computers how to speak. The results are truly impressive and online you find can examples of synthetic voices where the computer learns how to talk with the voice of celebrities such as Matt Damon. There are experiments showing that WaveNet improved the current state-of-the...