Upgrade to Pro — share decks privately, control downloads, hide ads and more …

文献紹介 8月10日

gumigumi7
August 10, 2018
120

文献紹介 8月10日

Improving Distributional Similarity
with Lessons Learned from Word Embeddings

gumigumi7

August 10, 2018
Tweet

Transcript

  1.  ▪  ▪ Omer Levy, Yoav Goldberg, Ido Dagan.

    ▪ Improving Distributional Similarity with Lessons Learned from Word Embeddings. ▪ Transactions of the Association for Computational Linguistics. 2015. ▪  ▪ , Skip-gram, LSA 2
  2.  ▪ , pi ur nu p i S a

    clulroug a S ▪ ) ▪ pi u tea , pi pi a a 3
  3.  ▪ V ▪ 2 count predict ▪ Don’t count,

    predict! (Baroni et al., 2014) ▪ Word2Vec Skip-gram Shifted-PMI ▪ Neural word embeddings as implicit matrix factorization (Levy and Goldberg, 2014) 4
  4.  ▪ ))( (- 2 ▪ )( D L AW

    P ▪ 2 - - 2 ▪ - - 2 - 2 VI M W ▪ GW))( (- 2 V M P ▪ ▪ 6
  5.  ▪ 2 ▪ C D 7 The quick brown

    fox jumps over the lazy dog 4 4 3 4 2 4 1 4 4 4 3 4 2 4 1 4 Word2Vec : 1 1 1 2 1 3 1 4 1 1 1 2 1 3 1 4 Glove :
  6.  ▪    ▪ '()&$,.-/+ ▪  #

     # % # ▪ !"$) 8 ! = 1 − % & &: , %: 
  7.  ▪ ▪ R G MR ▪ - - S

    ▪ Shifted PMI ▪ - 2 PID ▪ −log(&) W S NV 9 ( ) *( + = -./ 0, 2 − log(&) 3--./ 0, 2 = max(-./ 0, 2 − log(&), 0)
  8.  ▪ 7 0 7 577 2 ▪ 7 0.SW

    0 0 5 SV I ca db S W M b e P ▪ V S ▪ α WD C 10
  9.  ▪  ▪ PPMI-Matrix ▪ SVD ▪ SGNS (Word2Vec)

    ▪ Glove ▪  ▪   6$-"!#% ▪ &,-"  2$-"!#% ▪ '*(-")$+  ▪ 672)$+ 
  10.  ▪ - Se N ▪ - - - n

    y bc ▪ N smp n y i ▪ G tu d H uo a 18