Slide 8
Slide 8 text
LSTM Language Model
• LSTM Language Model:
Input word
(10000×1), one-hot
encoded
Embedded word (200×1)
or (650×1) , real-valued
Recurrent LSTM locks
Recurrent LSTM locks
Estimated next work
(10000×1), one-hot encoded
“give me that book” vs “give
that book to me”: which one
is better?
Small network: 200 LSTM blocks in each hidden layer.
Medium network: 650 LSTM blocks in each hidden layer.
Background Knowledge Poisoning of LSTM Language Model Experiments and Results Conclusions and Future Work