RNN and LSTM

Published: 09 Oct 2015 Category: deep_learning

Types of RNN

1) Plain Tanh Recurrent Nerual Networks

2) Gated Recurrent Neural Networks (GRU)

3) Long Short-Term Memory (LSTM)

Tutorials

A Beginner’s Guide to Recurrent Networks and LSTMs

http://deeplearning4j.org/lstm.html

A Deep Dive into Recurrent Neural Nets

http://nikhilbuduma.com/2015/01/11/a-deep-dive-into-recurrent-neural-networks/

Long Short-Term Memory: Tutorial on LSTM Recurrent Networks

http://people.idsia.ch/~juergen/lstm/index.htm

LSTM implementation explained

http://apaszke.github.io/lstm-explained.html

Recurrent Neural Networks Tutorial

Understanding LSTM Networks

Recurrent Neural Networks in DL4J

http://deeplearning4j.org/usingrnns.html

Train RNN

A Simple Way to Initialize Recurrent Networks of Rectified Linear Units

Sequence Level Training with Recurrent Neural Networks

Papers

Generating Sequences With Recurrent Neural Networks

DRAW: A Recurrent Neural Network For Image Generation

Unsupervised Learning of Video Representations using LSTMs(ICML2015)

LSTM: A Search Space Odyssey

Inferring Algorithmic Patterns with Stack-Augmented Recurrent Nets

A Critical Review of Recurrent Neural Networks for Sequence Learning

Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks(Winner of MSCOCO image captioning challenge, 2015)

Visualizing and Understanding Recurrent Networks(Andrej Karpathy, Justin Johnson, Fei-Fei Li)

Grid Long Short-Term Memory

Depth-Gated LSTM

Deep Knowledge Tracing

Alternative structures for character-level RNNs(INRIA & Facebook AI Research)

Pixel Recurrent Neural Networks (Google DeepMind)

Long Short-Term Memory-Networks for Machine Reading

Lipreading with Long Short-Term Memory

Associative Long Short-Term Memory

Representation of linguistic form and function in recurrent neural networks

Architectural Complexity Measures of Recurrent Neural Networks

Easy-First Dependency Parsing with Hierarchical Tree LSTMs

Training Input-Output Recurrent Neural Networks through Spectral Methods

Learn To Execute Programs

Learning to Execute

Neural Programmer-Interpreters (Google DeepMind)

A Programmer-Interpreter Neural Network Architecture for Prefrontal Cognitive Control

Convolutional RNN: an Enhanced Model for Extracting Features from Sequential Data

Attention Models

Recurrent Models of Visual Attention (Google DeepMind. NIPS2014)

Recurrent Model of Visual Attention(Google DeepMind)

Show, Attend and Tell: Neural Image Caption Generation with Visual Attention

A Neural Attention Model for Abstractive Sentence Summarization(EMNLP 2015. Facebook AI Research)

Effective Approaches to Attention-based Neural Machine Translation(EMNLP2015)

Generating Images from Captions with Attention

Attention and Memory in Deep Learning and NLP

Survey on the attention based RNN model and its applications in computer vision

Train RNN

Training Recurrent Neural Networks (PhD thesis)

Deep learning for control using augmented Hessian-free optimization


Hierarchical Conflict Propagation: Sequence Learning in a Recurrent Deep Neural Network

Recurrent Batch Normalization

Codes

NeuralTalk (Deprecated): a Python+numpy project for learning Multimodal Recurrent Neural Networks that describe images with sentences

NeuralTalk2: Efficient Image Captioning code in Torch, runs on GPU

char-rnn in Blocks

Project: pycaffe-recurrent

Using neural networks for password cracking

Recurrent neural networks for decoding CAPTCHAS

torch-rnn: Efficient, reusable RNNs and LSTMs for torch

Deploying a model trained with GPU in Torch into JavaScript, for everyone to use

Blog

Survey on Attention-based Models Applied in NLP

http://yanran.li/peppypapers/2015/10/07/survey-attention-model-1.html

Survey on Advanced Attention-based Models

http://yanran.li/peppypapers/2015/10/07/survey-attention-model-2.html

Online Representation Learning in Recurrent Neural Language Models

http://www.marekrei.com/blog/online-representation-learning-in-recurrent-neural-language-models/

Fun with Recurrent Neural Nets: One More Dive into CNTK and TensorFlow

http://esciencegroup.com/2016/03/04/fun-with-recurrent-neural-nets-one-more-dive-into-cntk-and-tensorflow/

Materials to understand LSTM

https://medium.com/@shiyan/materials-to-understand-lstm-34387d6454c1#.4mt3bzoau

Understanding LSTM and its diagrams (★★★★★)

Persistent RNNs: 30 times faster RNN layers at small mini-batch sizes (Greg Diamos, Baidu Silicon Valley AI Lab)

http://svail.github.io/persistent_rnns/

All of Recurrent Neural Networks

https://medium.com/@jianqiangma/all-about-recurrent-neural-networks-9e5ae2936f6e#.q4s02elqg

Resources

Awesome-rnn - A curated list of resources dedicated to RNN

http://jiwonkim.org/awesome-rnn/

Jürgen Schmidhuber’s page on Recurrent Neural Networks

http://people.idsia.ch/~juergen/rnn.html

Reading and Questions

Are there any Recurrent convolutional neural network network implementations out there ?