Upgrade to Pro — share decks privately, control downloads, hide ads and more …

EMNLP2018論文読み会:Deriving Machine Attention from ...

Avatar for Yuta Koreeda Yuta Koreeda
December 09, 2018

EMNLP2018論文読み会:Deriving Machine Attention from Human Rationales

Avatar for Yuta Koreeda

Yuta Koreeda

December 09, 2018
Tweet

More Decks by Yuta Koreeda

Other Decks in Research

Transcript

  1. Who am I? 是枝祐太 某電機会社リサーチャー 研究歴 • 〜2015: 医療+ロボット(大学) •

    〜2016: ロボット+応用機械学習 • 〜現在: 応用機械学習+自然言語処理 koreyou koreyou_ 2
  2. Who am I? 是枝祐太 某電機会社リサーチャー 研究歴 • 〜2015: 医療+ロボット(大学) •

    〜2016: ロボット+応用機械学習 • 〜現在: 応用機械学習+自然言語処理 koreyou koreyou_ 2
  3. tl;dr Bat et al. 2018. Deriving Machine Attention from Human

    Rationales. EMNLP. 目的:分類の根拠となった箇所のデータを用い 低リソースドメインで分類精度向上 手法:ドメイン非依存なRationale⇒attentionの 変換を学習 • Rationale:人間が作成した分類の根拠となる記載箇所 結果:観点付き評判分析の観点、ドメイン転移 でベースラインを上回った 3
  4. Deriving Machine Attention from Human Rationales Yujia Bao1, Shiyu Chang2,

    Mo Yu2, Regina Barzilay1 1Computer Science and Artif cial Intelligence Lab, MIT 2MIT-IBM Watson AI Lab, IBM Research {yujia, regina} @csail.mit.edu, [email protected], [email protected] 5
  5. 観点付き評判分析 観点付き評判分析をビークルに研究 本研究は自然言語タスク全般に利用可能 • わかりやすさのために具体的なタスクを先に紹介 観点付き評判分析 (Aspect-based sentiment analysis; ABSA)

    • 入力文が各観点について肯定的か否定的かを分類 本発表では“ドメイン”を“観点”と読みかえて理解 a nice and clean hotel to stay for business and leisure . but the location is not good if you need public transport . it took too long for transport and waiting for bus . but the swimming pool looks good . Location Cleanliness 6
  6. Rationale (根拠、解釈) 根拠 (Rationale) 提示型AIが注目されている Rationale=分類の根拠となる記載箇所 • なぜその予測をしたかの解釈を与えAIを説明可能にする Rationaleを提示する研究が注目されている[1, 2]

    a nice and clean hotel to stay for business and leisure . but the location is not good if you need public transport . it took too long for transport and waiting for bus . but the swimming pool looks good . Location Cleanliness Cleanliness Location 6
  7. 文分類におけるAttention機構の活用 Attention機構により文分類の精度向上が図れる プーリングとしてのattention機構 • 各単語表現からattentionの値 (実数) を計算 • attentionの値で単語表現の重み付き和 Task:

    Hotel location label: negative a nice and clean hotel to stay for business and leisure . but the location is not good if you need public transport . it took too long for transport and waiting for bus . but the swimming pool looks good . 8
  8. Attention vs. Rationale RationaleをAttention風に変換する Attention ̸= Rationale • Attentionは連続値(強弱がある)、rationaleは二値 •

    Attentionは分類精度を最大化するよう最適化されている Rationaleを直接学習に使うよりも、Rationaleを attention 風に変換してから学習に使うほうが良 いのでは?⇒R2A (rationale to attention) • 分類学習に適したAttentionを真(oracle) attentionと呼ぶ Task: Hotel location label: negative a nice and clean hotel to stay for business and leisure . but the location is not good if you need public transport . it took too long for transport and waiting for bus . but the swimming pool looks good . Attention Rationale 9
  9. 提案手法の流れ 文章 Attention生成 you get what you pay for .

    not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... 学習 分類器 R2A 学習 分類器 学習 Attention生成 you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... you get what you pay for . not the cleanest rooms but bed was clean and so was bath- room . bring your own towels though as very thin . service was excellent , ...... 文章 ラベル Rationale Rationale 真Attention ラベル 真Attention ターゲットドメイン(観点) ソース ドメイン(観点) 大量 少量 11
  10. 観点間の転移 提案手法の有効性を確認 Rationaleを加えると分類精度が向上 真attentionを介する提案手法では更に精度向上 • Attention ̸= Rationaleを確認 Source Target

    SVM RA-SVM‡ RA-CNN‡ TRANS† RA-TRANS‡† OURS‡† ORACLE† Beer aroma+palate Beer look 74.41 74.83 74.94 72.75 76.41 79.53 80.29 Beer look+palate Beer aroma 68.57 69.23 67.55 69.92 76.45 77.94 78.11 Beer look+aroma Beer palate 63.88 67.82 65.72 74.66 73.40 75.24 75.50 Table 3: Accuracy of transferring between aspects. Models with † use labeled data from source aspects. Models with ‡ use human rationales on the target aspect. Rationaleを学習に 使うベースライン Rationaleで attentionを 学習 Rationaleを 使わない 転移学習 ターゲット ドメインの 真attention を活用 14
  11. ドメイン間の転移 提案手法の有効性を確認 ドメイン間の転移でも同様の傾向を確認 ターゲットドメインの真attentionを使った Oracleは更に性能が高い Source Target SVM RA-SVM‡ RA-CNN‡

    TRANS† RA-TRANS‡† OURS‡† ORACLE† Beer look + Beer aroma + Beer palate Hotel location 78.65 79.09 79.28 80.42 82.10 84.52 85.43 Hotel cleanliness 86.44 86.68 89.01 86.95 87.15 90.66 92.09 Hotel service 85.34 86.61 87.91 87.37 86.40 89.93 92.42 Table 4: Accuracy of transferring between domains. Models with † use labeled data from source domains and unlabeled data from the target domain. Models with ‡ use human rationales on the target task. Rationaleを学習に 使うベースライン Rationaleで attentionを 学習 Rationaleを 使わない 転移学習 ターゲット ドメインの 真attention を活用 15
  12. 各機能の評価 Wasserstein距離を損失を入れたモデルで はhinvが共通の空間ある Rationaleよりも、R2Aで生成したattentionのほ うが真attentionに近い (a) OURS (b) OURS w/o

    L wd Figure 5: t-SNE visualization of the learned hidden representation5 for beer review (blue circle) and hotel review (orange triangle). Target Human rationales R2A-generated attention Location 0.5185 0.2371 Cleanliness 0.5948 0.3141 Service 0.5833 0.2871 Table 6: Avg. cosine distance to the oracle atten- tion over the target training set. The R2A is trained on beer reviews with unlabeled hotel reviews. 16
  13. Rationaleを使うことの効率性 データを増やすよりRationaleを与える方が効率的 Rationaleのデータを作るくらいならばデータを 増やせば良いのではないか? Rationaleであれば6.5%〜50%のデータ量で同等 の分類精度を得られる Accuracy 73.00 76.75 80.50

    84.25 88.00 Num. training examples 200 400 600 800 1000 1200 1400 1600 1800 2000 84.52 Ours
 (using 200) Attention-based classif er Accuracy 78.00 81.50 85.00 88.50 92.00 Num. training examples 200 700 1200 1700 2200 2700 3200 3700 90.66 Ours
 (using 200) Attention-based classif er Accuracy 80.00 83.00 86.00 89.00 92.00 Num. training examples 200 500 800 1100 1400 1700 2000 2300 89.93 Ours
 (using 200) Attention-based classif er Figure 7: Learning curve of an attention-based classif er on three tasks: hotel location (left), hotel clean- liness (center), hotel service (right). The performance of our approach trained on 200 examples with human rationales is shown as a reference. 17
  14. まとめ Bat et al. 2018. Deriving Machine Attention from Human

    Rationales. EMNLP. 目的:分類の根拠となった箇所のデータを用い 低リソースドメインで分類精度向上 手法:ドメイン非依存なRationale⇒attentionの 変換を学習 • Rationale:人間が作成した分類の根拠となる記載箇所 結果:観点付き評判分析の観点、ドメイン転移 でベースラインを上回った 18
  15. Get the presentation slides from: https://bit.ly/2Us6tFn https://github.com/koreyou/emnlp2018- meetup.git This presentation

    is licensed under CC0 1.0 Universal (CC0 1.0) Public Domain Dedication (except figures derived from the original paper[3]) . cz 20
  16. References i Tao Lei, Regina Barzilay, and Tommi Jaakkola. “Rationalizing

    Neural Predictions”. In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 2016, pp. 107–117. Wang Ling et al. “Program Induction by Rationale Generation: Learning to Solve and Explain Algebraic Word Problems”. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Vol. 1. 2017, pp. 158–167. Yujia Bao et al. “Deriving Machine Attention from Human Rationales”. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2018. url: http://www.aclweb.org/anthology/D18-1216. 21