HUGE SUCCESS
Speech, text understanding
Robotics / Computer Vision
Business / Big Data
Artificial General Intelligence (AGI)
Slide 18
Slide 18 text
How
its done ?
Slide 19
Slide 19 text
what why how next
Shallow Network
ℎ
ℎ = (, 0)
= ′(ℎ, 1)
= (, )
minimize
Slide 20
Slide 20 text
what why how next
Deep Network
Slide 21
Slide 21 text
what why how next
Deep Network
More abstract features
Stellar performance
Vanishing Gradient
Overfitting
Slide 22
Slide 22 text
what why how next
Autoencoder
ℎ
Unsupervised Feature
Learning
Slide 23
Slide 23 text
what why how next
Stacked Autoencoder
Y. Bengio et. all; Greedy Layer-Wise Training of Deep Networks
Slide 24
Slide 24 text
what why how next
Stacked Autoencoder
1. Unsupervised,
layer by layer pretraining
2. Supervised fine tuning
Slide 25
Slide 25 text
what why how next
Deep Belief Network
2006 breakthrough
Stacking Restricted Boltzmann Machines (RBMs)
Hinton, G. E., Osindero, S. and Teh, Y.; A fast learning algorithm for deep belief nets
Slide 26
Slide 26 text
Rethinking
Computer
Vision
Slide 27
Slide 27 text
what why how next
Traditional
Image Classification pipeline
Feature Extraction
(SIFT, SURF etc.)
Classifier
(SVM, NN etc.)
Slide 28
Slide 28 text
what why how next
Convolutional Neural Network
Images taken from deeplearning.net
Slide 29
Slide 29 text
what why how next
Convolutional Neural Network
Slide 30
Slide 30 text
what why how next
Convolutional Neural Network
Images taken from deeplearning.net
Slide 31
Slide 31 text
what why how next
Convolutional Neural Network
Slide 32
Slide 32 text
what why how next
The Starry Night
Vincent van Gogh
Leon A. Gatys, Alexander S. Ecker and Matthias Bethge; A Neural Algorithm of Artistic Style
Slide 33
Slide 33 text
what why how next
Slide 34
Slide 34 text
what why how next
Scene Description
CNN + RNN
Oriol Vinyals et. all; Show and Tell: A Neural Image Caption Generator
Slide 35
Slide 35 text
Learning
Sequences
Slide 36
Slide 36 text
what why how next
Recurrent Neural Network
Simple Elman Version
ℎ
ℎ
= (
, ℎ−1
, 0, 1)
= ′(ℎ
, 2)
Slide 37
Slide 37 text
what why how next
Long Short Term Memory (LSTM)
add memory cells
learn access mechanism
Sepp Hochreiter and Jürgen Schmidhuber; Long short-term memory
Slide 38
Slide 38 text
No content
Slide 39
Slide 39 text
what why how next
Slide 40
Slide 40 text
what why how next
Fooling Deep Networks
Anh Nguyen, Jason Yosinski, Jeff Clune; Deep Neural Networks are Easily Fooled
Slide 41
Slide 41 text
Next
Cool things
to try
Slide 42
Slide 42 text
what why how next
Hyperparameter optimization
bayesian
Optimization methods
adadelta, rmsprop . . .
Regularization
dropout, dither . . .
Slide 43
Slide 43 text
what why how next
Attention & Memory
NTMs, Memory Networks, Stack RNNs . . .
NLP
Translation, description
Slide 44
Slide 44 text
what why how next
Cognitive Hardware
FPGA, GPU, Neuromorphic Chips
Scalable DL
map-reduce, compute clusters
Slide 45
Slide 45 text
what why how next
Deep Reinforcement Learning
deepmindish things, deep Q learning
Energy models
RBMs, DBNs . . .
Slide 46
Slide 46 text
https://www.reddit.com/r/MachineLearning/wiki
Slide 47
Slide 47 text
Theano (Python) | Torch (lua) | Caffe (C++)
Github is a friend