$30 off During Our Annual Pro Sale. View Details »
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
セミパラメトリック推論の基礎の復習
Search
Daisuke Yoneoka
November 14, 2023
Research
0
120
セミパラメトリック推論の基礎の復習
Daisuke Yoneoka
November 14, 2023
Tweet
Share
More Decks by Daisuke Yoneoka
See All by Daisuke Yoneoka
感染症の数理モデル15
kingqwert
0
12
感染症の数理モデル14
kingqwert
0
110
感染症の数理モデル13
kingqwert
0
40
感染症の数理モデル12
kingqwert
0
110
感染症の数理モデル11
kingqwert
0
110
感染症の数理セミナー_10_.pdf
kingqwert
0
130
感染症の数理モデル9
kingqwert
0
100
感染症の数理モデル8
kingqwert
0
110
感染症の数理モデル7
kingqwert
0
110
Other Decks in Research
See All in Research
Multi-Agent Large Language Models for Code Intelligence: Opportunities, Challenges, and Research Directions
fatemeh_fard
0
120
LLM-jp-3 and beyond: Training Large Language Models
odashi
1
730
Pythonでジオを使い倒そう! 〜それとFOSS4G Hiroshima 2026のご紹介を少し〜
wata909
0
1.2k
言語モデルの地図:確率分布と情報幾何による類似性の可視化
shimosan
8
2.4k
データサイエンティストをめぐる環境の違い2025年版〈一般ビジネスパーソン調査の国際比較〉
datascientistsociety
PRO
0
340
[論文紹介] Intuitive Fine-Tuning
ryou0634
0
160
その推薦システムの評価指標、ユーザーの感覚とズレてるかも
kuri8ive
1
290
生成的情報検索時代におけるAI利用と認知バイアス
trycycle
PRO
0
120
Self-Hosted WebAssembly Runtime for Runtime-Neutral Checkpoint/Restore in Edge–Cloud Continuum
chikuwait
0
230
CoRL2025速報
rpc
3
3.7k
第二言語習得研究における 明示的・暗示的知識の再検討:この分類は何に役に立つか,何に役に立たないか
tam07pb915
0
500
[Devfest Incheon 2025] 모두를 위한 친절한 언어모델(LLM) 학습 가이드
beomi
2
1.3k
Featured
See All Featured
16th Malabo Montpellier Forum Presentation
akademiya2063
PRO
0
31
Statistics for Hackers
jakevdp
799
230k
Unsuck your backbone
ammeep
671
58k
A better future with KSS
kneath
240
18k
Writing Fast Ruby
sferik
630
62k
B2B Lead Gen: Tactics, Traps & Triumph
marketingsoph
0
33
Exploring anti-patterns in Rails
aemeredith
2
210
Lessons Learnt from Crawling 1000+ Websites
charlesmeaden
PRO
0
960
Future Trends and Review - Lecture 12 - Web Technologies (1019888BNR)
signer
PRO
0
3.1k
The Organizational Zoo: Understanding Human Behavior Agility Through Metaphoric Constructive Conversations (based on the works of Arthur Shelley, Ph.D)
kimpetersen
PRO
0
200
[RailsConf 2023] Rails as a piece of cake
palkan
58
6.2k
The World Runs on Bad Software
bkeepers
PRO
72
12k
Transcript
ηϛύϥϝτϦοΫਪͷجૅͷ෮श Daisuke Yoneoka September 29, 2014
Notations جຊతʹ Tsiatis,2006 ʹै͏. Θ͔Μͳ͔ͬͨΒࣗͰௐͯͶ! ϕΫτϧߦྻଠࣈʹͯ͠ͳ͍͚Ͳ, ͦࣗ͜Ͱิ͍ͬͯͩ͘͞. σʔλ i.i.d Ͱ
Zi = (Zi1, . . . , Zim) ∈ Rm αϯϓϧαΠζ n ਓ. i.e., Z1, . . . , Zn φ(Z) Өڹؔ u(Zi, θ) ਪఆؔ Լ͖ࣈͷ eff (ۙ) ༗ޮ (efficient) ͱ͍͏ҙຯ
ηϛύϥϝτϦοΫਪͱʁ Zi ͷີ͕ؔηϛύϥϝτϦοΫϞσϧʹै͏ͱ S = {p(z : θ, η)|θ ∈
Θ ⊂ Rr, η ∈ H} θ ༗ݶ࣍ݩͷڵຯ͋ΔύϥϝλͰ, η ແݶ࣍ݩͷͲ͏Ͱ͍͍ύ ϥϝλ (ہ֎ (nuisance) ύϥϝʔλʔ). ηϛύϥϝτϦοΫਪ: ͜ͷͱͰ θ ͷ࠷ྑͷਪఆྔ (RAL ਪఆ ྔ) ΛͱΊΔ͜ͱ
Өڹؔ θ ͳΜͰ͍͍͔Β࠷ྑΛݟ͚ͭΔͱ͍͏ͷແཧήʔ → Ϋϥε Λݶఆͯͦ͜͠Ͱݟ͚ͭΔ! (౷ܭͰΑ͘ΔΑͶ) Өڹؔ: ਪఆྔ ˆ
θ ͷӨڹؔͱ, (Ϟʔϝϯτʹ੍͕͋Δ) √ n(ˆ θ − θ) = 1 √ n n i=1 φ(Zi, θ, η) + op(1) Λຬͨ͢ϕΫτϧؔ. ˆ θ ۙઢܗਪఆྔͱݺͼ n → ∞ ͰҰகੑ ͱۙਖ਼نੑ͕͋Δ √ n(ˆ θ − θ) → N 0, E[φ(Zi, θ, η)φ(Zi, θ, η)T ] Πϝʔδతʹ͋Δσʔλ͕ͲΕ͚ͩਪఆʹӨڹΛ༩͍͑ͯΔ͔Λ දݱͨ͠ͷ
ਪఆؔͱ M ਪఆ ਪఆํఔࣜ n i=1 u(Zi, θ) ਪఆؔ =
0 ͷղͱͯ͠ಘΒΕΔͷΛ M ਪఆྔ ͱݺͿ. Α͘ݟΔ score ؔͳΜ͔ίϨ. ͨͩ͠, E[φ(Zi, θ)] = 0 ظ 0 , E[∥φ(Zi, θ)∥2] < ∞ ࢄతͳͷൃࢄ͠ͳ͍ . ͋ͱ͏গ͚ͩ݅͋͠Δ. Ұகੑͱۙਖ਼نੑΛ࣋ͭ √ n(ˆ θ − θ) = 1 √ n n i=1 E[ ∂u(Zi, θ) ∂θ ] −1 u(Zi, θ) ͕͜͜Өڹؔʹͳ͍ͬͯΔ +op(1) → N 0, E[ ∂u(Zi, θ) ∂θ ] −1 E[u(Zi, θ)u(Zi, θ)T ] E[ ∂u(Zi, θ) ∂θ ] −T ] ͜ͷۙࢄͷਪఆྔΛαϯυΠονਪఆྔͱݺΜͩΓ͢Δ
RAL ਪఆྔ ۙઢܥਪఆྔͳΜ͔ྑͦ͞͏ʂͰ super efficiency ͷ (Hodges) ͕Δʂ Super efficiency:
ۙతʹ Cramer-Rao ͷԼݶΑΓྑ͍ͷ͕Ͱ͖ Δͷ͜ͱ ͜ͷΛղܾͨ͠ͷ͕ RAL (Regular asymptotic linear) ਪఆྔ. ͦͷਖ਼ଇ݅ۃݶ͕ LDGP (local data generating process) ʹґ ଘ͠ͳ͍͜ͱ (ৄ͘͠ Tsiatis, 2006) ηϛύϥਪ͜ͷ RAL ਪఆྔͷӨڹؔΛٻΊΔ͜ͱΛߟ͑Δ
Parametric submodel ηϛύϥϝτϦοΫϞσϧ S ͷ֤ʹର͠ p(z; θ, η) ∈ Ssub
⊂ S Λຬͨ͢ύϥϝτϦοΫϞσϧ Ssub = {p(z; θ, γ)|θ ∈ Θ ⊂ Rr, γ ∈ Γ ⊂ Rs, s ∈ N} ΛύϥϝτϦοΫαϒϞσϧͱݺͿ.
Nuisance tangent space (ہ֎ۭؒ) ηϛύϥϝτϦοΫϞσϧ S ͷ֤ʹର͠, ύϥϝτϦοΫαϒϞσϧ Ssub ͷہ֎ۭؒΛ
TN θ,γ (Ssub) = {BT sγ(z, θ, γ)|B ∈ Rs} ͱ͢Δ. γ p(z; θ, η) ʹରԠ͢ΔͷͰ sγ(z, θ, γ) = ∂ ∂γ log p(z; θ, γ) Ͱ ද͞ΕΔ nuisance score ؔ. ͜ͷઢܗۭؒ͜ͷ nuisance score vector ʹ ΑͬͯுΒΕ͍ͯΔ. ͜ͷͱ͖ TN θ,η (S) = Ssub TN θ,γ (Ssub) Λ S ্ͷ p(z; θ, η) ʹ͓͚Δہ֎ۭؒͱΑͿ. ͪͳΈʹ, ଆͷू ߹ʹؔͯ͠ closure ΛͱΔԋࢉࢠ. Note:͜ͷۭؒେͰޙʹ, RAL ਪఆྔͷӨڹؔ͜ͷۭؒʹަۭͨؒ͠ʹ ଐ͢Δ͜ͱ͕ॏཁʹͳͬͯ͘Δʂ
ઢܗ෦ۭؒͷࣹӨͷزԿͱϐλΰϥεͷఆཧ
RAL ਪఆྔͷӨڹؔͷॏཁͳఆཧ ηϛύϥϝτϦοΫ RAL ਪఆྔ β ͷӨڹؔ φ(Z) ҎԼͷ݅Λຬ ͢Δ.
Corollary1 E[φ(Z)sβ] = E[φ(Z)sT efficient (Z, β0, η0)] = I. ͨͩ͠, s είΞؔͰ, sT efficient ༗ޮείΞؔ Corollary2 φ(Z) ہ֎ۭؒʹަ͍ͯ͠Δ. ༗ޮӨڹ্ؔͷ 2 ͭͷ݅Λຬͨ͠, ͦͷࢄߦྻ, ޮݶքΛୡ ͦ͠Ε φeffi(Z, β0, η0) = E[seff (Z, β0, η0)sT eff (Z, β0, η0)] −1 seff (Z, β0, η0)
ηϛύϥۭؒͷఆཧ ύϥϝτϦοΫαϒϞσϧͷ߹ͷ RAL ਪఆྔͷӨڹؔͱۭؒͱͷؔ Tsiatis, 2006 ͷ Ch4.3 ͋ͨΓΛݟͯͶʂ ఆཧ
1 RAL ਪఆྔͷӨڹؔ {φ(Z) + TN θ,η (S)⊥} ͱ͍͏ۭؒʹؚ·ΕΔ. ͨͩ͠, φ(Z) ҙͷ RAL ਪఆྔͷӨڹؔͰ, TN θ,η (S)⊥ ηϛύϥϝτϦο Ϋۭؒͷަิۭؒ ఆཧ 2 ηϛύϥϝτϦοΫ༗ޮͳਪఆྔ, ͦͷӨڹ͕ؔҰҙʹ well-defined Ͱܾఆ͞ Ε,φefficient = φ(Z) − {φ(Z)|TN θ,η (S)⊥} ͷཁૉ. ͪͳΈʹ, (h|U) projection of h ∈ H(ੵΛಋೖͨ͠ώϧϕϧτۭؒ) onto the space U (ઢܗۭؒ)
GEE ʹ͍ͭͯͷ Remarks Liang-Zeger ͷ GEE ͷηϛύϥϝτϦοΫϞσϧ (੍ϞʔϝϯτϞσϧ: 1 ࣍ͱ
2 ࣍ͷϞʔϝϯτʹ੍͚ͩΛஔ͍ͨϞσϧ) ҎԼͷಛΛͭ. ہॴ (ۙ༗) ޮਪఆྔ: ࢄؔͷԾఆ͕ਖ਼͚͠Ε, ༗ޮਪఆྔ Robustness: ແݶ࣍ݩͷύϥϝʔλਪఆ͕ඞཁ͕ͩ, ࢄؔΛ misspecify ͨ͠ͱͯ͠Ұகੑͱۙਖ਼نੑอ࣋ GEE ͷຊΛಡΊΘ͔Δ͚Ͳ, Working covariance matrix Λؒҧ͑ͯ ༗ޮੑࣦΘΕΔ͕, ͦͷଞͷ·͍͠ੑ࣭ (ۙਖ਼نੑͱҰகੑ) อ࣋Ͱ͖Δͬͯ͜ͱ