Long Short-Term Memory (LSTM) is a Recurrent Neural Network (RNN) architecture that looks at a sequence and remembers values over long intervals. LSTMs have been known to have achieved state of the art performance in many sequence classification problems. In this talk, I’ll cover how to write an LSTM using TensorFlow’s Python API for natural language understanding. This is going to be a code-heavy talk where I will implement the LSTM model and explain the math behind it step-by-step.
In short, it will cover:
– Understanding how the math behind LSTM architecture works in case of sequence classification
– Writing an LSTM model using TensorFlow for sentiment classification of variable length English language sentences
– Gotchas of using LSTMs on real data