Introduction to Deep Learning
We will discuss about the topics you liked to revisit (Mainly RNNs/LSTMs and Attention). Please:
Here are the Reading Assignments of RNNs/LSTMs and about Attention
Deep Learning Book - Chapter 10: Sequence Modeling: Recurrent and Recursive Nets, Sections 10.6, 10.7 and 10.10 to 10.12
For an overview on the topic, start with the blog post “Attention and Memory in Deep Learning and NLP” by Denny Britz and the distill.pub article “Attention and Augmented Recurrent Neural Networks” by Olah & Carter.
Continue by reading the following articles that provide details on two very recently proposed approaches,
Ashish et al. “Attention Is All You Need” in NIPS (2017) (The version on arxiv seems to be most recent.)
Follow the links from the blog post to the referenced articles for further reading.