This is a short companion page to our internal reading group of the book “Hands-on Machine Learning with Scikit-Learn, Keras, and TensorFlow, 2nd Edition”. However I unashamedly used a lot of PyTorch examples.
Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, 2nd Edition
Chapter 15. Processing Sequences Using RNNs and CNNs The batter hits the ball. The outfielder immediately starts running, anticipating the ball's trajectory. He tracks it, adapts his movements, and finally ... - Selection from Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, 2nd Edition [Book]
Understanding LSTM Networks
Posted on August 27, 2015 Humans don't start their thinking from scratch every second. As you read this essay, you understand each word based on your understanding of previous words. You don't throw everything away and start thinking from scratch again. Your thoughts have persistence.
- one to many: image -> caption sentence
- many to one: sentence -> sentiment (positive / negative label)
- many to many: a sentence in English -> a sentence in Turkish
- the other many to many: frames of video -> coordinates of bounding boxes around an object
Visualizing A Neural Machine Translation Model (Mechanics of Seq2seq Models With Attention)
Translations: Chinese (Simplified), Japanese, Korean, Russian, Turkish Watch: MIT's Deep Learning State of the Art lecture referencing this post May 25th update: New graphics (RNN animation, word embedding graph), color coding, elaborated on the final attention example. Note: The animations below are videos.
WaveNet: A Generative Model for Raw Audio
This post presents WaveNet, a deep generative model of raw audio waveforms. We show that WaveNets are able to generate speech which mimics any human voice and which sounds more natural than the best existing Text-to-Speech systems, reducing the gap with human performance by over 50%.