Upgrade to PRO for Only $50/Year—Limited-Time Offer! 🔥
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
An introduction of statistical learning
Search
Nakamura, Ryotaro
June 28, 2017
Science
0
43
An introduction of statistical learning
Nakamura, Ryotaro
June 28, 2017
Tweet
Share
More Decks by Nakamura, Ryotaro
See All by Nakamura, Ryotaro
Duct for beginners.
nryotaro
0
4.2k
Learn Go in 15 minutes
nryotaro
0
42
Seven architectural patterns
nryotaro
1
110
Improving Performance with Parallel Programming
nryotaro
0
55
Other Decks in Science
See All in Science
データベース10: 拡張実体関連モデル
trycycle
PRO
0
1k
HajimetenoLT vol.17
hashimoto_kei
1
110
データマイニング - コミュニティ発見
trycycle
PRO
0
180
データベース06: SQL (3/3) 副問い合わせ
trycycle
PRO
1
700
AIに仕事を奪われる 最初の医師たちへ
ikora128
0
1k
データマイニング - ノードの中心性
trycycle
PRO
0
320
機械学習 - 授業概要
trycycle
PRO
0
280
Lean4による汎化誤差評価の形式化
milano0017
1
380
データベース12: 正規化(2/2) - データ従属性に基づく正規化
trycycle
PRO
0
1k
機械学習 - pandas入門
trycycle
PRO
0
380
白金鉱業Meetup_Vol.20 効果検証ことはじめ / Introduction to Impact Evaluation
brainpadpr
2
1.4k
DMMにおけるABテスト検証設計の工夫
xc6da
1
1.4k
Featured
See All Featured
Writing Fast Ruby
sferik
630
62k
[SF Ruby Conf 2025] Rails X
palkan
0
490
実際に使うSQLの書き方 徹底解説 / pgcon21j-tutorial
soudai
PRO
196
69k
Unsuck your backbone
ammeep
671
58k
Site-Speed That Sticks
csswizardry
13
990
Testing 201, or: Great Expectations
jmmastey
46
7.8k
The Straight Up "How To Draw Better" Workshop
denniskardys
239
140k
Large-scale JavaScript Application Architecture
addyosmani
515
110k
Designing Dashboards & Data Visualisations in Web Apps
destraynor
231
54k
Improving Core Web Vitals using Speculation Rules API
sergeychernyshev
21
1.3k
Optimizing for Happiness
mojombo
379
70k
Product Roadmaps are Hard
iamctodd
PRO
55
12k
Transcript
ػցֶशษڧձ ୈ 1 ճ தଜ ྒྷଠ June 13, 2017
Table of contents Supervised Learning 1. Classification 2. Perceptron 3.
Regression Unsupervised Learning 4. Clustering 1
ࠓͷඪ ࣍ճҎ߱ʹֶͿΞϧΰϦζϜͷ֓ཁΛΔ ΞϧΰϦζϜͱద༻ྫ ΞϧΰϦζϜ ద༻ྫ ྨ εύϜϝʔϧఆ ճؼੳ ച্༧ଌ ΫϥελϦϯά
ը૾ͷݮ৭ॲཧ 2
ύϥϝτϦοΫ๏ ϞσϧʢࣜʣΛԾఆ͠ɼϞσϧͷ࠷దͳύϥϝλΛֶश͢Δ ύϥϝτϦοΫ๏ͷखॱ 1. σʔλͷ༧ଌϞσϧΛԾఆ 2. Ϟσϧͷύϥϝλͷ ධՁج४ΛܾΊΔ 3. ύϥϝλΛܾΊΔ
0.0 0.2 0.4 0.6 0.8 1.0 −0.2 0.0 0.2 0.4 0.6 0.8 1.0 Ұ࣍ؔͷϞσϧͷύϥϝλௐ 3
Classification
ྨ Ϋϥεʹྨ͞ΕͨطଘσʔλΛݩʹ৽نσʔλΛྨ͢Δ ΞϧΰϦζϜ • ύʔηϓτϩϯ • ϩδεςΟοΫճؼ 4
Perceptron
ύʔηϓτϩϯ, Ϟσϧ ઢܗͳϞσϧ f Λઃఆ͢Δ f (x, y) = w0
+ w1x + w2y f (x, y) > 0 ⇒ t = +1 f (x, y) < 0 ⇒ t = −1 −20 −10 0 10 20 30 x −30 −20 −10 0 10 20 y t = +1 t = -1 ଐੑ t = ±1 Λͭσʔλ܈ ઢ্ͷ (x′, y′) f (x′, y′) = 0 ΛΈͨ͢ 5
ύʔηϓτϩϯ, ධՁج४ʢޡࠩؔʣ ޡࠩؔ E ͕࠷খʹͳΔ wi ΛٻΊΔ E = N
∑ i=1 {− (w0 + w1x + w2y) ti } = N ∑ i=1 (−f (xi , yi )ti ) • N σʔλ • ޡྨͩͱ −f (xi , yi )ti > 0 −20 −10 0 10 20 30 x −30 −20 −10 0 10 20 y t = +1 t = -1 ଐੑ t = ±1 Λͭσʔλ܈ 6
ϩδεςΟοΫճؼ, Ϟσϧ ύʔηϓτϩϯͱಉ͘͡ઢܗϞσϧ f Λઃఆ͢Δ f (x, y) = w0
+ w1x + w2y f (x, y) > 0 ⇒ t = +1 f (x, y) < 0 ⇒ t = −1 −30 −20 −10 0 10 20 30 x −20 −15 −10 −5 0 5 10 15 20 y t = +1 t = -1 f (x, y) ͕૿Ճ͢Δ͖ 7
ϩδεςΟοΫճؼ, Ϟσϧ ͨͩ͠ɼ|f | ͕େ͖͍΄Ͳ t Ͱ͋Δ͕֬ߴ͍ͱ͢Δ ϩδεςΟοΫؔ σ (α)
= 1 1 + e−α Λಋೖ͠ɼ (x′, y′) ͕ t = 1 Ͱ͋Δ֬Λ 0 < σ ( f ( x′, y′ )) < 1 ͱ͢Δ −4 −3 −2 −1 0 1 2 3 4 α 0.0 0.2 0.4 0.6 0.8 1.0 σ (α) ϩδεςΟοΫؔͷάϥϑ 8
ϩδεςΟοΫճؼ, ධՁج४ʢ࠷ਪఆʣ ܇࿅σʔλ͕ಘΒΕΔ֬ P Λ࠷େʹ͢Δ wi ΛٻΊΔ p(x, y) =
σ(x0 + w1x + w2y) P = N ∏ i p (xi , yi )tn {1 − p (xi , yi )}1−tn ܇࿅σʔλ࠷ൃੜ͕֬ߴ͍σʔλͰ͋ΔͱԾఆ͍ͯ͠Δ 9
Regression
ճؼੳ, ϞσϧͱධՁج४ʢ࠷খೋʣ σʔλ͕ M ࣍ଟ߲ࣜ f ʹै͏ͱͯ͠ɼೋޡࠩ ED Λ࠷খʹ͢Δ ύϥϝλ
wi ΛબͿ f (x) = M ∑ m=0 wmxm ED = 1 2 N ∑ n=1 {f (xn) − tn}2 0 2 4 6 8 10 −15 −10 −5 0 5 ground truth degree 3 degree 4 degree 5 training points M ∈ {3, 4, 5} ͷଟ߲ࣜۙࣅྫ 10
Clustering
k ฏۉ๏ σʔλؒͷڑΛٻΊɼσʔλΛ k ݸͷΫϥελʹ͚Δ −2 −1 0 1 2
3 0 1 2 3 4 5 σʔλू߹ −2 −1 0 1 2 3 0 1 2 3 4 5 cluster 1 cluster 2 cluster 3 centroids k = 3 ͷΫϥελ Ϋϥελ͝ͱʹදσʔλΛܾΊɼදͷۙ͘ͷσʔλू߹Ͱ ΫϥελΛ࡞Δ 11
k ฏۉ๏ͷΞϧΰϦζϜ ೖྗ: σʔλू߹ D = { x(1), x(2), ·
· · , x(|D|) } : Ϋϥελ k ແ࡞ҝʹ m1, m2 · · · , mk ΛܾΊΔ until ऩଋ foreach x(i) ∈ D cmax = arg max c sim ( x(i), mc ) σʔλू߹ͷׂ insert x(i)into cmax end foreach ∀c, mc = 1 |c| ∑ x(i)∈c x(i) දϕΫτϧΛ࠶ܭࢉ end until 12