Upgrade to Pro
— share decks privately, control downloads, hide ads and more …
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
Gradient Descent Easy
Search
soulchild
July 23, 2014
Science
0
83
Gradient Descent Easy
Easy/brief version of Gradient Descent from Artificial Intelligence Lecture
soulchild
July 23, 2014
Tweet
Share
More Decks by soulchild
See All by soulchild
Similarities between macOS and iOS development
soulchild
0
120
N Tier Architecture for MMORPG
soulchild
0
93
Other Decks in Science
See All in Science
AI(人工知能)の過去・現在・未来 —AIは人間を超えるのか—
tagtag
PRO
0
140
サイコロで理解する原子核崩壊と拡散現象 〜単純化されたモデルで本質を理解する〜
syotasasaki593876
0
150
データマイニング - コミュニティ発見
trycycle
PRO
0
200
データベース11: 正規化(1/2) - 望ましくない関係スキーマ
trycycle
PRO
0
1k
会社でMLモデルを作るとは @電気通信大学 データアントレプレナーフェロープログラム
yuto16
1
520
LayerXにおける業務の完全自動運転化に向けたAI技術活用事例 / layerx-ai-jsai2025
shimacos
4
21k
主成分分析に基づく教師なし特徴抽出法を用いたコラーゲン-グリコサミノグリカンメッシュの遺伝子発現への影響
tagtag
PRO
0
180
知能とはなにかーヒトとAIのあいだー
tagtag
PRO
0
170
データマイニング - ノードの中心性
trycycle
PRO
0
330
baseballrによるMLBデータの抽出と階層ベイズモデルによる打率の推定 / TokyoR118
dropout009
2
660
Distributional Regression
tackyas
0
340
MCMCのR-hatは分散分析である
moricup
0
590
Featured
See All Featured
B2B Lead Gen: Tactics, Traps & Triumph
marketingsoph
0
53
A brief & incomplete history of UX Design for the World Wide Web: 1989–2019
jct
1
300
How to Build an AI Search Optimization Roadmap - Criteria and Steps to Take #SEOIRL
aleyda
1
1.9k
Automating Front-end Workflow
addyosmani
1371
200k
Reflections from 52 weeks, 52 projects
jeffersonlam
356
21k
Navigating Team Friction
lara
192
16k
SEO for Brand Visibility & Recognition
aleyda
0
4.2k
A Modern Web Designer's Workflow
chriscoyier
698
190k
Stop Working from a Prison Cell
hatefulcrawdad
273
21k
Let's Do A Bunch of Simple Stuff to Make Websites Faster
chriscoyier
508
140k
Paper Plane (Part 1)
katiecoart
PRO
0
4.2k
The Success of Rails: Ensuring Growth for the Next 100 Years
eileencodes
47
7.9k
Transcript
Artificial Intelligence Gradient Descent soulchild
Gradient Descent Let computer find the minimum point in a
given graph or equation
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 Minimum point
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 Minimum point How does a computer know that this is a minimum point?
Gradient Descent Ans : By brute-forcing the derivative until a
value equal or near to 0 is found y = x2 dy dx = 2x Then guess x by starting from, eg: -6 to 6
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4 dy dx = 0
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4 dy dx = 0 Minimum point found, stop
Gradient Descent A better way to brute force
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent How to set a good learning rate (α)
?
Gradient Descent When to stop searching ?
Gradient Descent When to stop searching ? Set a maximum
number of iteration
Gradient Descent When to stop searching ? Set a maximum
number of iteration dy dx < n When n can be 0.1, 0.01,.. etc
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0 then computer stop finding
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0 then computer stop finding What about this?! smaller than previous point wor
Gradient Descent Weakness of Gradient Descent Gradient Descent may stuck
in a local minima thus can’t find the global minima
Gradient Descent Q&A