Upgrade to Pro
— share decks privately, control downloads, hide ads and more …
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
Gradient Descent Easy
Search
soulchild
July 23, 2014
Science
0
83
Gradient Descent Easy
Easy/brief version of Gradient Descent from Artificial Intelligence Lecture
soulchild
July 23, 2014
Tweet
Share
More Decks by soulchild
See All by soulchild
Similarities between macOS and iOS development
soulchild
0
120
N Tier Architecture for MMORPG
soulchild
0
93
Other Decks in Science
See All in Science
AIによる科学の加速: 各領域での革新と共創の未来
masayamoriofficial
0
410
20251212_LT忘年会_データサイエンス枠_新川.pdf
shinpsan
0
230
academist Prize 4期生 研究トーク延長戦!「美は世界を救う」っていうけど、どうやって?
jimpe_hitsuwari
0
470
検索と推論タスクに関する論文の紹介
ynakano
1
140
【RSJ2025】PAMIQ Core: リアルタイム継続学習のための⾮同期推論・学習フレームワーク
gesonanko
0
640
機械学習 - ニューラルネットワーク入門
trycycle
PRO
0
940
Text-to-SQLの既存の評価指標を問い直す
gotalab555
1
170
機械学習 - K-means & 階層的クラスタリング
trycycle
PRO
0
1.2k
baseballrによるMLBデータの抽出と階層ベイズモデルによる打率の推定 / TokyoR118
dropout009
2
660
Cross-Media Technologies, Information Science and Human-Information Interaction
signer
PRO
3
32k
Optimization of the Tournament Format for the Nationwide High School Kyudo Competition in Japan
konakalab
0
140
データマイニング - ウェブとグラフ
trycycle
PRO
0
240
Featured
See All Featured
Designing for humans not robots
tammielis
254
26k
No one is an island. Learnings from fostering a developers community.
thoeni
21
3.6k
Measuring Dark Social's Impact On Conversion and Attribution
stephenakadiri
1
120
職位にかかわらず全員がリーダーシップを発揮するチーム作り / Building a team where everyone can demonstrate leadership regardless of position
madoxten
57
50k
Designing Experiences People Love
moore
144
24k
The Art of Programming - Codeland 2020
erikaheidi
57
14k
Testing 201, or: Great Expectations
jmmastey
46
8k
Thoughts on Productivity
jonyablonski
74
5k
Deep Space Network (abreviated)
tonyrice
0
48
GraphQLの誤解/rethinking-graphql
sonatard
74
11k
Put a Button on it: Removing Barriers to Going Fast.
kastner
60
4.2k
Design of three-dimensional binary manipulators for pick-and-place task avoiding obstacles (IECON2024)
konakalab
0
350
Transcript
Artificial Intelligence Gradient Descent soulchild
Gradient Descent Let computer find the minimum point in a
given graph or equation
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 Minimum point
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 Minimum point How does a computer know that this is a minimum point?
Gradient Descent Ans : By brute-forcing the derivative until a
value equal or near to 0 is found y = x2 dy dx = 2x Then guess x by starting from, eg: -6 to 6
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4 dy dx = 0
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 dy dx = -12 dy dx = —8 dy dx = -4 dy dx = 0 Minimum point found, stop
Gradient Descent A better way to brute force
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent y 0 10 20 30 40 x -6
-4 -2 0 2 4 6 y = x2 In short, it works like this
Gradient Descent How to set a good learning rate (α)
?
Gradient Descent When to stop searching ?
Gradient Descent When to stop searching ? Set a maximum
number of iteration
Gradient Descent When to stop searching ? Set a maximum
number of iteration dy dx < n When n can be 0.1, 0.01,.. etc
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0 then computer stop finding
Gradient Descent Weakness of Gradient Descent 0 3 6 9
12 -8 -6 -4 -2 0 2 4 let say start from here dy dx = 0 then computer stop finding What about this?! smaller than previous point wor
Gradient Descent Weakness of Gradient Descent Gradient Descent may stuck
in a local minima thus can’t find the global minima
Gradient Descent Q&A