Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Modern Techniques for Dimensional Reduction

techsessions
December 07, 2017

Modern Techniques for Dimensional Reduction

Takeaways:
– What is dimensional reduction and how/why you should use it;
– How t-SNE and PCA work as two distinct ML methods;
– Witness some amazing results using t-SNE!

techsessions

December 07, 2017
Tweet

More Decks by techsessions

Other Decks in Technology

Transcript

  1. D A T A S C I E N C

    E Modern Techniques for Dimensional Reduction 1 Copyright © ASI 2017 All rights reserved Machine Learning Tech Sessions 7th December 2017 Ilya Feige Head of ML Research
  2. Copyright © ASI 2017 All rights reserved • Head of

    Research Ilya Feige 2 Copyright © ASI 2017 All rights reserved
  3. Copyright © ASI 2017 All rights reserved About ASI Data

    Science 4 TECHNOLOGY PEOPLE TRAINING EXPERTISE CONSULTING Copyright © ASI 2017 All rights reserved
  4. Copyright © ASI 2017 All rights reserved Outline 5 1.

    Dimensional reduction motivation 2. PCA reminder 3. T-SNE is amazing! 5
  5. Copyright © ASI 2017 All rights reserved Dimensional Reduction 6

    The process of reducing the number of variables (features) under consideration in a statistical / machine learning analysis
  6. Copyright © ASI 2017 All rights reserved Why is it

    needed? 7 • Nowadays data is very high dimensional • Many features are highly correlated • Manual feature selection is often impossible • Visualising data requires 2D representations • Most models break down in high dimensions! 1-D Data occupies ⇠ ✓ 1 10 ◆2 of space Data occupies ⇠ ✓ 1 10 ◆n of space Data occupies ⇠ 1 10 of space 2-D n-D Curse of dimensionality
  7. Copyright © ASI 2017 All rights reserved Principal Component Analysis

    (PCA) 9 Idea: Find basis that better expresses data How: Eigenvalue decomposition (or SVD) What for: Can then throw away directions of low variance • Rotate to uncorrelated coordinates • Project onto largest variance axes Dimensional reduction with PCA:
  8. Copyright © ASI 2017 All rights reserved 10 MNIST is

    a dataset images of handwritten digits
  9. Copyright © ASI 2017 All rights reserved Definition of t-SNE

    13 “t-distributed stochastic neighbour embedding” Step 1: Construct a distribution in the high- dimensional space based on pair-wise distance Step 2: Construct a similar distribution (but with wider tails) in the low-dimensional space Step 3: Make the two distributions as similar as possible by minimising their KL divergence pj|i = exp ||xi xj ||2/2 2 i P k6=i exp ||xi xk ||2/2 2 i pij = 1 2N pi|j + pj|i qij = 1 + ||yi yj ||2 1 P k6=` 1 + ||yk y` ||2 1 {y⇤ i } = argmin yi n KL P||Q o = argmin yi ⇢ X j6=k pjk log pjk qjk 1 2 3
  10. Copyright © ASI 2017 All rights reserved 14 t-SNE is

    able to separate MNIST digits incredibly well
  11. Copyright © ASI 2017 All rights reserved 16 PCA effectively

    separates data in low dimensions PCA on Genotypes: