Upgrade to Pro — share decks privately, control downloads, hide ads and more …

OS@IBM - The Machine in Sheep’s Clothing: Trust and Transparency in the ML Lifecycle

OS@IBM - The Machine in Sheep’s Clothing: Trust and Transparency in the ML Lifecycle

Tools/Communities:
AI Fairness 360 Toolkit: http://aif360.mybluemix.net/
Model Asset Exchange: http://ibm.biz/model-exchange
IBM's Center for Open Source Data and AI Technologies: http://ibm.biz/codait-projects
IBM's Internal Call for Code Competition: https://w3.ibm.com/developer/callforcode/
IBM's Public Call for Code Competition: https://callforcode.org/

Talk Sources:

Podcasts/Tweets
https://leanin.org/podcast-episodes/siri-is-artificial-intelligence-biased
https://art19.com/shows/the-ezra-klein-show/episodes/663fd0b7-ee60-4e3e-b2cb-4fcb4040eef1
https://twitter.com/alexisohanian/status/1087973027055316994

Amazon
https://www.aclu.org/blog/privacy-technology/surveillance-technologies/amazons-face-recognition-falsely-matched-28
https://www.openmic.org/news/2019/1/16/halt-rekognition

Google
https://motherboard.vice.com/en_us/article/j5jmj8/google-artificial-intelligence-bias

COMPAS
https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
https://www.technologyreview.com/s/612775/algorithms-criminal-justice-ai/

Data for Black Lives
http://d4bl.org/about.html
2019 Conference Notes: https://docs.google.com/document/d/1E1mfgTp73QFRmNBunl8cIpyUmDos28rekidux0voTsg/edit?ts=5c39f92e

Gender Shades Project
http://news.mit.edu/2018/study-finds-gender-skin-type-bias-artificial-intelligence-systems-0212
https://www.youtube.com/watch?time_continue=1&v=TWWsW1w-BVo
https://www.ajlunited.org/fight

Other resources referenced in this talk:
https://www.nytimes.com/2018/02/12/business/computer-science-ethics-courses.html
https://www.vox.com/science-and-health/2017/4/17/15322378/how-artificial-intelligence-learns-how-to-be-racist
https://www.engadget.com/2019/01/24/pinterest-skin-tone-search-diversity/

Maureen McElaney

May 08, 2019
Tweet

More Decks by Maureen McElaney

Other Decks in Technology

Transcript

  1. Open Source @ IBM Blueprint Talk The Machine in Sheep’s

    Clothing Trust and Transparency in the ML Lifecycle May 8th, 2019
  2. Open Source @ IBM May 8, 2019 / © 2019

    IBM Corporation 2 JEFFREY BOREK WW Program Director, Open Technology & IP Mgt. IBM Cognitive Applications [email protected] @jeffborek
  3. The Machine in Sheep’s Clothing Trust and Transparency in the

    ML Lifecycle Maureen McElaney Developer Advocate IBM Center for Open Source Data and AI Technologies [email protected] @Mo_Mack May 8, 2019 / © 2019 IBM Corporation
  4. The Machine in Sheep’s Clothing Building Trust and Transparency into

    the ML Lifecycle 4 May 8, 2019 / © 2019 IBM Corporation
  5. “A cognitive bias is a systematic pattern of deviation from

    norm or rationality in judgment. Individuals create their own "subjective social reality" from their perception of the input.” - Wikipedia 6
  6. October 2017 - Google Natural Language API https://cloud.google.com/natural-language/ 10 Source:

    https://motherboard.vice.com/en_us/article/j5jmj8/google-artificial-intelligence-bias
  7. October 2017 - Google Natural Language API https://cloud.google.com/natural-language/ 11 Source:

    https://motherboard.vice.com/en_us/article/j5jmj8/google-artificial-intelligence-bias
  8. October 2017 - Google Natural Language API https://cloud.google.com/natural-language/ 12 Source:

    https://motherboard.vice.com/en_us/article/j5jmj8/google-artificial-intelligence-bias
  9. “We will correct this specific case, and, more broadly, building

    more inclusive algorithms is crucial to bringing the benefits of machine learning to everyone.” 13
  10. BLACK VS. WHITE DEFENDANTS ◦ Falsely labeled black defendants as

    likely of future crime at twice the rate as white defendants. ◦ White defendants mislabeled as low risk more than black defendants ◦ Pegged Black defendants 77% more likely to be at risk of committing future violent crime 21
  11. “If we fail to make ethical and inclusive artificial intelligence

    we risk losing gains made in civil rights and gender equity under the guise of machine neutrality.” 28 - Joy Boulamwini @jovialjoy
  12. Solutions? What can we do to combat bias in AI?

    29 May 8, 2019 / © 2019 IBM Corporation
  13. Questions posed to students in these courses... Is the technology

    fair? How do you make sure that the data is not biased? Should machines be judging humans? 34 May 8, 2019 / © 2019 IBM Corporation
  14. “By combining the latest in machine learning and inclusive product

    development, we're able to directly respond to Pinner feedback and build a more useful product.” 39 - Candice Morgan @Candice_MMorgan
  15. Tool #1: AI Fairness 360 Toolkit Open Source Library 41

    May 8, 2019 / © 2019 IBM Corporation
  16. TYPES OF METRICS ◦ Individual vs. Group Fairness, or Both

    ◦ Group Fairness: Data vs Model ◦ Group Fairness: We’re All Equal vs What You See is What You Get ◦ Group Fairness: Ratios vs Differences 44
  17. Machine Learning Pipeline In- Processing Pre- Processing Post- Processing 47

    Modifying the training data. Modifying the learning algorithm. Modifying the predictions (or outcomes.)
  18. Tool #2: Model Asset Exchange Open Source Pre-Trained Deep Learning

    Models 52 May 8, 2019 / © 2019 IBM Corporation
  19. Step 1: Find a model ...that does what you need

    ...that is free to use ...that is performant enough 53
  20. Step 2: Get the code Is there a good implementation

    available? ...that does what you need ...that is free to use ...that is performant enough 54
  21. Step 3: Verify the model ◦ Does it do what

    you need? ◦ Is it free to use (license)? ◦ Is it performant enough? ◦ Accuracy? 55
  22. Step 5: Deploy your model ◦ Adjust inference code (or

    write from scratch) ◦ Package inference code, model code, and pre-trained weights together ◦ Deploy your package 58
  23. Model Asset Exchange The Model Asset Exchange (MAX) is a

    one stop shop for developers/data scientists to find and use free and open source deep learning models ibm.biz/model-exchange 60
  24. ◦ Wide variety of domains (text, audio, images, etc) ◦

    Multiple deep learning frameworks ◦ Vetted and tested code/IP ◦ Build and deploy a model web service in seconds 61 Model Asset Exchange
  25. Dedicated Open Source Efforts IBM Center for Open-Source Data and

    AI Technologies (CODAIT) 64 May 8, 2019 / © 2019 IBM Corporation
  26. Center for Open Source Data and AI Technologies CODAIT aims

    to make AI solutions dramatically easier to create, deploy, and manage in the enterprise. 40 open source developers! 65 CODAIT codait.org May 8, 2019 / © 2019 IBM Corporation
  27. 67 Or is it? That’s a lot of open source

    developers! May 8, 2019 / © 2019 IBM Corporation
  28. Active IBM Contributors to Open Source (Committed code on Github

    in 2018) 68 May 8, 2019 / © 2019 IBM Corporation
  29. Active IBM Users of Open Source (Certified to consume and/or

    contribute open source in 2018) 69 May 8, 2019 / © 2019 IBM Corporation
  30. 74 Photo by rawpixel on Unsplash No matter what it

    is our responsibility to build systems that are fair.