Upgrade to Pro
— share decks privately, control downloads, hide ads and more …
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
An introduction of statistical learning
Search
Sponsored
·
Your Podcast. Everywhere. Effortlessly.
Share. Educate. Inspire. Entertain. You do you. We'll handle the rest.
→
Nakamura, Ryotaro
June 28, 2017
Science
0
44
An introduction of statistical learning
Nakamura, Ryotaro
June 28, 2017
Tweet
Share
More Decks by Nakamura, Ryotaro
See All by Nakamura, Ryotaro
Duct for beginners.
nryotaro
0
4.2k
Learn Go in 15 minutes
nryotaro
0
43
Seven architectural patterns
nryotaro
1
110
Improving Performance with Parallel Programming
nryotaro
0
58
Other Decks in Science
See All in Science
Distributional Regression
tackyas
0
420
コンピュータビジョンによるロボットの視覚と判断:宇宙空間での適応と課題
hf149
1
580
(2025) Balade en cyclotomie
mansuy
0
520
コミュニティサイエンスの実践@日本認知科学会2025
hayataka88
0
150
20251212_LT忘年会_データサイエンス枠_新川.pdf
shinpsan
0
260
Lean4による汎化誤差評価の形式化
milano0017
1
460
検索と推論タスクに関する論文の紹介
ynakano
1
180
データベース15: ビッグデータ時代のデータベース
trycycle
PRO
0
470
因果推論と機械学習
sshimizu2006
1
1k
Bear-safety-running
akirun_run
0
120
DMMにおけるABテスト検証設計の工夫
xc6da
1
1.6k
見上公一.pdf
genomethica
0
110
Featured
See All Featured
The Psychology of Web Performance [Beyond Tellerrand 2023]
tammyeverts
49
3.3k
The Myth of the Modular Monolith - Day 2 Keynote - Rails World 2024
eileencodes
26
3.4k
10 Git Anti Patterns You Should be Aware of
lemiorhan
PRO
659
61k
The Cost Of JavaScript in 2023
addyosmani
55
9.8k
Bridging the Design Gap: How Collaborative Modelling removes blockers to flow between stakeholders and teams @FastFlow conf
baasie
0
490
JAMstack: Web Apps at Ludicrous Speed - All Things Open 2022
reverentgeek
1
400
Self-Hosted WebAssembly Runtime for Runtime-Neutral Checkpoint/Restore in Edge–Cloud Continuum
chikuwait
0
420
From π to Pie charts
rasagy
0
160
"I'm Feeling Lucky" - Building Great Search Experiences for Today's Users (#IAC19)
danielanewman
231
22k
We Analyzed 250 Million AI Search Results: Here's What I Found
joshbly
1
1.1k
Building AI with AI
inesmontani
PRO
1
820
Easily Structure & Communicate Ideas using Wireframe
afnizarnur
194
17k
Transcript
ػցֶशษڧձ ୈ 1 ճ தଜ ྒྷଠ June 13, 2017
Table of contents Supervised Learning 1. Classification 2. Perceptron 3.
Regression Unsupervised Learning 4. Clustering 1
ࠓͷඪ ࣍ճҎ߱ʹֶͿΞϧΰϦζϜͷ֓ཁΛΔ ΞϧΰϦζϜͱద༻ྫ ΞϧΰϦζϜ ద༻ྫ ྨ εύϜϝʔϧఆ ճؼੳ ച্༧ଌ ΫϥελϦϯά
ը૾ͷݮ৭ॲཧ 2
ύϥϝτϦοΫ๏ ϞσϧʢࣜʣΛԾఆ͠ɼϞσϧͷ࠷దͳύϥϝλΛֶश͢Δ ύϥϝτϦοΫ๏ͷखॱ 1. σʔλͷ༧ଌϞσϧΛԾఆ 2. Ϟσϧͷύϥϝλͷ ධՁج४ΛܾΊΔ 3. ύϥϝλΛܾΊΔ
0.0 0.2 0.4 0.6 0.8 1.0 −0.2 0.0 0.2 0.4 0.6 0.8 1.0 Ұ࣍ؔͷϞσϧͷύϥϝλௐ 3
Classification
ྨ Ϋϥεʹྨ͞ΕͨطଘσʔλΛݩʹ৽نσʔλΛྨ͢Δ ΞϧΰϦζϜ • ύʔηϓτϩϯ • ϩδεςΟοΫճؼ 4
Perceptron
ύʔηϓτϩϯ, Ϟσϧ ઢܗͳϞσϧ f Λઃఆ͢Δ f (x, y) = w0
+ w1x + w2y f (x, y) > 0 ⇒ t = +1 f (x, y) < 0 ⇒ t = −1 −20 −10 0 10 20 30 x −30 −20 −10 0 10 20 y t = +1 t = -1 ଐੑ t = ±1 Λͭσʔλ܈ ઢ্ͷ (x′, y′) f (x′, y′) = 0 ΛΈͨ͢ 5
ύʔηϓτϩϯ, ධՁج४ʢޡࠩؔʣ ޡࠩؔ E ͕࠷খʹͳΔ wi ΛٻΊΔ E = N
∑ i=1 {− (w0 + w1x + w2y) ti } = N ∑ i=1 (−f (xi , yi )ti ) • N σʔλ • ޡྨͩͱ −f (xi , yi )ti > 0 −20 −10 0 10 20 30 x −30 −20 −10 0 10 20 y t = +1 t = -1 ଐੑ t = ±1 Λͭσʔλ܈ 6
ϩδεςΟοΫճؼ, Ϟσϧ ύʔηϓτϩϯͱಉ͘͡ઢܗϞσϧ f Λઃఆ͢Δ f (x, y) = w0
+ w1x + w2y f (x, y) > 0 ⇒ t = +1 f (x, y) < 0 ⇒ t = −1 −30 −20 −10 0 10 20 30 x −20 −15 −10 −5 0 5 10 15 20 y t = +1 t = -1 f (x, y) ͕૿Ճ͢Δ͖ 7
ϩδεςΟοΫճؼ, Ϟσϧ ͨͩ͠ɼ|f | ͕େ͖͍΄Ͳ t Ͱ͋Δ͕֬ߴ͍ͱ͢Δ ϩδεςΟοΫؔ σ (α)
= 1 1 + e−α Λಋೖ͠ɼ (x′, y′) ͕ t = 1 Ͱ͋Δ֬Λ 0 < σ ( f ( x′, y′ )) < 1 ͱ͢Δ −4 −3 −2 −1 0 1 2 3 4 α 0.0 0.2 0.4 0.6 0.8 1.0 σ (α) ϩδεςΟοΫؔͷάϥϑ 8
ϩδεςΟοΫճؼ, ධՁج४ʢ࠷ਪఆʣ ܇࿅σʔλ͕ಘΒΕΔ֬ P Λ࠷େʹ͢Δ wi ΛٻΊΔ p(x, y) =
σ(x0 + w1x + w2y) P = N ∏ i p (xi , yi )tn {1 − p (xi , yi )}1−tn ܇࿅σʔλ࠷ൃੜ͕֬ߴ͍σʔλͰ͋ΔͱԾఆ͍ͯ͠Δ 9
Regression
ճؼੳ, ϞσϧͱධՁج४ʢ࠷খೋʣ σʔλ͕ M ࣍ଟ߲ࣜ f ʹै͏ͱͯ͠ɼೋޡࠩ ED Λ࠷খʹ͢Δ ύϥϝλ
wi ΛબͿ f (x) = M ∑ m=0 wmxm ED = 1 2 N ∑ n=1 {f (xn) − tn}2 0 2 4 6 8 10 −15 −10 −5 0 5 ground truth degree 3 degree 4 degree 5 training points M ∈ {3, 4, 5} ͷଟ߲ࣜۙࣅྫ 10
Clustering
k ฏۉ๏ σʔλؒͷڑΛٻΊɼσʔλΛ k ݸͷΫϥελʹ͚Δ −2 −1 0 1 2
3 0 1 2 3 4 5 σʔλू߹ −2 −1 0 1 2 3 0 1 2 3 4 5 cluster 1 cluster 2 cluster 3 centroids k = 3 ͷΫϥελ Ϋϥελ͝ͱʹදσʔλΛܾΊɼදͷۙ͘ͷσʔλू߹Ͱ ΫϥελΛ࡞Δ 11
k ฏۉ๏ͷΞϧΰϦζϜ ೖྗ: σʔλू߹ D = { x(1), x(2), ·
· · , x(|D|) } : Ϋϥελ k ແ࡞ҝʹ m1, m2 · · · , mk ΛܾΊΔ until ऩଋ foreach x(i) ∈ D cmax = arg max c sim ( x(i), mc ) σʔλू߹ͷׂ insert x(i)into cmax end foreach ∀c, mc = 1 |c| ∑ x(i)∈c x(i) දϕΫτϧΛ࠶ܭࢉ end until 12