Upgrade to Pro — share decks privately, control downloads, hide ads and more …

MACHINE LEARNING INTERPRETABILITY: WHY AND HOW!

OmaymaS
December 10, 2019
360

MACHINE LEARNING INTERPRETABILITY: WHY AND HOW!

With the increasing adoption of machine learning-based solutions in different domains, systems that use black-box algorithms are getting used more often with the promise of providing higher accuracy. However, this accuracy comes at the cost of interpretability, which introduces a barrier against wider adoption of such algorithms in crucial areas and raises the skepticism of the impacted individuals. This talk focuses on the importance of interpretable machine learning, why it is crucial from technical and ethical perspectives and its current limitations. In addition, it gives an overview of some of the relevant tools and packages (e.g. LIME, SHAPLEY).

OmaymaS

December 10, 2019
Tweet

Transcript

  1. • You will be given two labels with the corresponding

    mapping to 0/1. • You will be asked to classify some images. • You are a classifier. HOW WOULD YOUR MENTAL MODEL LABEL THIS IMAGE? RULES
  2. CAT → 0 DOG → 1 OR HOW WOULD YOUR

    MENTAL MODEL LABEL THIS IMAGE?
  3. CAT → 0 DOG → 1 OR HOW WOULD YOUR

    MENTAL MODEL LABEL THIS IMAGE?
  4. DUCK → 0 RABBIT → 1 OR HOW WOULD YOUR

    MENTAL MODEL LABEL THIS IMAGE?
  5. “Uses AI to give you more insight into candidates, so

    you can make better decisions.” $#&T NONSENSE Source: Business Insider Video (2017) 25000 FEATURES → INSIGHT SCORE
  6. COLLECT/LABEL DATA IT IS HUMANS WHO BIAS IN: - REPRESENTATION

    - DISTRIBUTION - LABELS AND MORE….. WRITE ALGORITHMS DEFINE METRICS
  7. IT IS HUMANS WHO DEFINE METRICS WRITE ALGORITHMS COLLECT/LABEL DATA

    - TRAIN/TEST SPLIT - FEATURES/PROXIES - MODEL COMPLEXITY AND MORE…..
  8. IT IS HUMANS WHO COLLECT/LABEL DATA DEFINE METRICS WRITE ALGORITHMS

    - WHAT IS THE IMPACT OF DIFFERENT ERROR TYPES ON DIFFERENT GROUPS? - WHAT DO YOU OPTIMIZE FOR?
  9. PRACTITIONERS CONSISTENTLY: - OVERESTIMATE THEIR MODEL’S ACCURACY. - PROPAGATE FEEDBACK

    LOOPS. - FAIL TO NOTICE DATA LEAKS. “ ” “Why Should I Trust You?” Explaining the Predictions of Any Classifier
  10. 1- Select a point to explain (red). Based on an

    example in “Interpretable Machine Learning” Book by Christoph Molnar LIME (Tabular Data)
  11. 2- Sample data points. LIME (Tabular Data) Based on an

    example in “Interpretable Machine Learning” Book by Christoph Molnar
  12. 3- Weight points according to their proximity to the selected

    point. LIME (Tabular Data) Based on an example in “Interpretable Machine Learning” Book by Christoph Molnar
  13. 4- Train a weighted, interpretable local model. LIME (Tabular Data)

    Based on an example in “Interpretable Machine Learning” Book by Christoph Molnar
  14. 5- Explain the black-box model prediction using the local model.

    LIME (Tabular Data) Based on an example in “Interpretable Machine Learning” Book by Christoph Molnar
  15. Label: tabby, tabby cat Probability: 0.29 Explanation Fit: 0.77 Label:

    Egyptian Cat Probability: 0.28 Explanation Fit: 0.69 LIME (Images) Original Model: pre-trained ImageNet model
  16. Label: tabby, tabby cat Probability: 0.29 Explanation Fit: 0.77 Label:

    Egyptian Cat Probability: 0.28 Explanation Fit: 0.69 Type: Supports Type: Contradicts LIME (Images) Original Model: pre-trained ImageNet model
  17. LIME (Images) “Why Should I Trust You?” Explaining the Predictions

    of Any Classifier https:/ /arxiv.org/pdf/1602.04938.pdf
  18. Boring Stupid Dumb Waste information Label predicted: -ve sentiment LIME

    (Text) Original Model: Keras model (CNN+LSTM)
  19. Pros LIME - Provides human-friendly explanations. - Gives a fidelity

    measure. - Can use other features than the black-box model.
  20. Pros LIME Cons - Provides human-friendly explanations. - Gives a

    fidelity measure. - Can use other features than the original model. - The definition of proximity is not totally resolved in tabular data. - Instability of explanations.
  21. Pros - Provides human-friendly explanations. - Gives a fidelity measure.

    - Can use other features than the original model. Cons - Instability of explanations. LIME - The definition of proximity is not totally resolved in tabular data.
  22. SHAPLEY VALUES Explain the difference between the actual prediction and

    the average/baseline prediction of the black-box model. coalitional game theory “ ”
  23. Pros - Solid theory. - The difference between the prediction

    and the average prediction is fairly distributed among the feature values of the instance. SHAPLEY VALUES
  24. Pros Cons - Solid theory - The difference between the

    prediction and the average prediction is fairly distributed among the feature values of the instance. - Computationally expensive. - Can be misinterpreted. SHAPLEY VALUES - Uses all features (not ideal for explanations that contain few features).
  25. Pros Cons - Solid theory. - The difference between the

    prediction and the average prediction is fairly distributed among the feature values of the instance. - Computationally expensive. - Can be misinterpreted. SHAPLEY VALUES - Uses all features (not ideal for explanations that contain few features).
  26. Explainable AI for Trees: From Local Explanations to Global Understanding

    Bar chart (left) and SHAP summary plot (right) for a gradient boosted decision tree model trained on the mortality dataset. SHAP
  27. • Deploy the model. • Log feature attributions for certain

    predictions (e.g. rare class). • Monitor model training/live skew. • Monitor attribution skew and focus on the most influential features. PREDICTION AUDITING AND MODEL MONITORING Explainable AI for Trees: From Local Explanations to Global Understanding Google AI Explainability Whitepaper
  28. https:/ /cloud.google.com/ml-engine/docs/ai-explanations/limitations • Explanations are LOCAL (Each attribution only shows

    how much the feature affected the prediction for that particular example). • Explanations/Feature attributions are subject to adversarial attacks as predictions in complex models. • Explanations alone cannot tell if your model is fair, unbiased, or of sound quality • Different methods are just complementary tools to be combined with other approaches and the practitioners’ best judgement. • Explanations might be misinterpreted in some cases. LIMITATIONS Google AI Explainability Whitepaper Limitations of Interpretable Machine Learning Methods