Upgrade to Pro
— share decks privately, control downloads, hide ads and more …
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
Gradient Descent Easy
Search
soulchild
July 23, 2014
Science
83
0
Share
Gradient Descent Easy
Easy/brief version of Gradient Descent from Artificial Intelligence Lecture
soulchild
July 23, 2014
More Decks by soulchild
See All by soulchild
Similarities between macOS and iOS development
soulchild
0
130
N Tier Architecture for MMORPG
soulchild
0
95
Other Decks in Science
See All in Science
NDCG is NOT All I Need
statditto
2
3k
白金鉱業Vol.21【初学者向け発表枠】身近な例から学ぶ数理最適化の基礎 / Learning the Basics of Mathematical Optimization Through Everyday Examples
brainpadpr
1
680
機械学習 - ニューラルネットワーク入門
trycycle
PRO
0
970
人生を変えた一冊「独学大全」のはなし / Self-study ENCYCLOPEDIA: The Book Which Change My Life #独学大全 #EM推し本
expajp
0
140
【論文紹介】Is CLIP ideal? No. Can we fix it?Yes! 第65回 コンピュータビジョン勉強会@関東
shun6211
5
2.4k
Navigating Weather and Climate Data
rabernat
0
150
データベース11: 正規化(1/2) - 望ましくない関係スキーマ
trycycle
PRO
0
1.1k
ド文系だった私が、 KaggleのNCAAコンペでソロ金取れるまで
wakamatsu_takumu
2
2.2k
Bear-safety-running
akirun_run
0
120
HDC tutorial
michielstock
1
590
俺たちは本当に分かり合えるのか? ~ PdMとスクラムチームの “ずれ” を科学する
bonotake
2
2.1k
Rashomon at the Sound: Reconstructing all possible paleoearthquake histories in the Puget Lowland through topological search
cossatot
0
790
Featured
See All Featured
Exploring anti-patterns in Rails
aemeredith
2
300
The Cult of Friendly URLs
andyhume
79
6.8k
New Earth Scene 8
popppiees
2
1.9k
Building a Scalable Design System with Sketch
lauravandoore
463
34k
Understanding Cognitive Biases in Performance Measurement
bluesmoon
32
2.8k
A brief & incomplete history of UX Design for the World Wide Web: 1989–2019
jct
1
330
Writing Fast Ruby
sferik
630
63k
Sharpening the Axe: The Primacy of Toolmaking
bcantrill
46
2.7k
Optimising Largest Contentful Paint
csswizardry
37
3.6k
How to make the Groovebox
asonas
2
2.1k
Building Applications with DynamoDB
mza
96
7k
The Limits of Empathy - UXLibs8
cassininazir
1
280
Transcript
Artificial Intelligence Gradient Descent soulchild
Gradient Descent Let computer find the minimum point in a
given graph or equation
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 Minimum point
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 Minimum point How does a computer know that this is a minimum point?
Gradient Descent Ans : By brute-forcing the derivative until a
value equal or near to 0 is found y = x2 dy dx = 2x Then guess x by starting from, eg: -6 to 6
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4 dy dx = 0
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4 dy dx = 0 Minimum point found, stop
Gradient Descent A better way to brute force
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent How to set a good learning rate (α)
?
Gradient Descent When to stop searching ?
Gradient Descent When to stop searching ? Set a maximum
number of iteration
Gradient Descent When to stop searching ? Set a maximum
number of iteration dy dx < n When n can be 0.1, 0.01,.. etc
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0 then computer stop finding
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0 then computer stop finding What about this?! smaller than previous point wor
Gradient Descent Weakness of Gradient Descent Gradient Descent may stuck
in a local minima thus can’t find the global minima
Gradient Descent Q&A