Slide 1

Slide 1 text

Prediction and Control: Watching Algorithms Jennifer Helsby (@redshiftzero) Postdoctoral Fellow Computation Institute University of Chicago 32c3 December 29, 2015

Slide 2

Slide 2 text

• Algorithmic landscape • Advantages of intelligent systems • Current applications • The nightmare case • Properties of algorithmic systems • Fairness • Transparency • Accountability • Paths forward Outline

Slide 3

Slide 3 text

Technology is neither good not bad; nor is it neutral. - Kranzberg’s First Law of Technology

Slide 4

Slide 4 text

government collection social media IoT devices, sensors medical history voice location genetic information biometrics browsing profiles wearables emails and messaging logs

Slide 5

Slide 5 text

• slow • biased • not transparent • difficult to audit • fast • unbiased? • transparent? • can audit? human decision algorithmic decision

Slide 6

Slide 6 text

Data Inference Action

Slide 7

Slide 7 text

Algorithmic decision making is expanding from industry into new domains.

Slide 8

Slide 8 text

No content

Slide 9

Slide 9 text

No content

Slide 10

Slide 10 text

No content

Slide 11

Slide 11 text

VOTE VOTE VOTE VOTE VOTE VOTE VOTE VOTE

Slide 12

Slide 12 text

data-driven city

Slide 13

Slide 13 text

Advantages • Automate rote tasks • Enable us to distill huge volumes of data down to the critical pieces of information • Optimize resource allocation • Optimize actions to produce desired outcomes

Slide 14

Slide 14 text

Algorithms can have serious implications.

Slide 15

Slide 15 text

Chinese Social Credit System ranks each citizen based on their behavior to be used by all citizens by 2020 • financial record • criminal history • drivers license history • some medical information • purchase history • social media monitoring • social network analysis • … “carry forward sincerity and traditional virtues”

Slide 16

Slide 16 text

What do we want?

Slide 17

Slide 17 text

Privacy Fairness Transparency Accountability What do we want?

Slide 18

Slide 18 text

I. Fairness

Slide 19

Slide 19 text

Predictive Policing

Slide 20

Slide 20 text

Predictive Policing individual level geographic level

Slide 21

Slide 21 text

No content

Slide 22

Slide 22 text

No content

Slide 23

Slide 23 text

Individual Fairness similar people treated similarly Group Fairness protected groups treated similarly Fairness in Errors errors should not be concentrated in protected groups Zemel et al. 2013, Dwork et al. 2011 Fairness Concerns

Slide 24

Slide 24 text

Training Data Issues • crime databases contain only crimes detected by police • effect of biased policing, e.g. black people arrested at rates 3.73x that of whites for marijuana crimes [ACLU]

Slide 25

Slide 25 text

• Race and location are correlated in the US: machine learning systems can learn sensitive features • No opt-out: citizens need more information about these systems

Slide 26

Slide 26 text

II. Transparency

Slide 27

Slide 27 text

What are these systems doing? ?

Slide 28

Slide 28 text

Complexity Algorithms must be human interpretable

Slide 29

Slide 29 text

Influence

Slide 30

Slide 30 text

Alerting users to the fact manipulation might be occurring did not decrease the amplitude of the effect. Conservative Conservative Liberal Liberal Control Control Manipulate Manipulate Epstein and Robertson 2015 “it is a relatively simple matter to mask the bias in search rankings so that it is undetectable to virtually every user” Influencing Voting Introduced 20% shift in voter preferences

Slide 31

Slide 31 text

behaviouralinsights.co.uk

Slide 32

Slide 32 text

behaviouralinsights.co.uk

Slide 33

Slide 33 text

No content

Slide 34

Slide 34 text

Optimization Industry: • Time spent on website • Click through rate • Likes • Profit Politics: • Votes for the desired candidate Policy: • Better use of government services • Voter registration • Health outcomes • Education outcomes • Compliance

Slide 35

Slide 35 text

The invisible barbed wire of big data limits our lives to a space that might look quiet and enticing enough, but is not of our own choosing and that we cannot rebuild or expand. The worst part is that we do not see it as such. Because we believe that we are free to go anywhere, the barbed wire remains invisible. -Morozov

Slide 36

Slide 36 text

III. Accountability

Slide 37

Slide 37 text

oversight

Slide 38

Slide 38 text

No content

Slide 39

Slide 39 text

Greene vs. San Francisco • San Francisco ALPR gets (false) hit on car • Traffic stop of 47-year old woman • Officers conducted pat-down, search, held at gunpoint

Slide 40

Slide 40 text

Auditing

Slide 41

Slide 41 text

? Inputs: Test accounts Real accounts (collaborative approach) Outputs: Was one output shown to one user and not another?

Slide 42

Slide 42 text

? Browsing profile Ad 1 Ad 2 … Ad n

Slide 43

Slide 43 text

Datta et al 2015 Ad Network Browsing Profile Browsing Profile Browsing Profile Browsing Profile Treatment 2 Treatment 1 Ad 1 … Ad n Ad 1 … Ad n Ad 1 … Ad n Ad 1 … Ad n

Slide 44

Slide 44 text

No content

Slide 45

Slide 45 text

Who is in control?

Slide 46

Slide 46 text

Audit tools Sunlight: https://columbia.github.io/sunlight/ AdFisher: https://github.com/tadatitam/info-flow-experiments/ Anonymity systems (Tor) View the landscape Obfuscate by injecting noise (Adnauseam.io) Technology

Slide 47

Slide 47 text

Policy Regulation Independent ethics review 3rd party audits

Slide 48

Slide 48 text

Closing thoughts Algorithmic systems need to have appropriate oversight in order to be controlled Hacking and privacy advocacy community has an important role to play in this fight Thanks! Jennifer Helsby @redshiftzero Postdoctoral Fellow Computation Institute University of Chicago

Slide 49

Slide 49 text

References ACLU, “The War on Marijuana”, https://www.aclu.org/files/assets/aclu- thewaronmarijuana-rel2.pdf Bond, Farris, Jones, Kramer, Marlow, Settle and Fowler, “A 61-milion-person experiment in social influence and political mobilization”, Nature 489, 295–298 , 2012 Datta, Tschantz, and Datta, “Automated Experiments on Ad Privacy Settings”, Proceedings on Privacy Enhancing Technologies 2015; 2015 (1):92-112 Dwork, Hardt, Pitassi, Reingold, and Zemel, “Fairness through Awareness”, ITCS '12 Proceedings of the 3rd Innovations in Theoretical Computer Science Conference Epstein and Robertson, “The search engine manipulation effect (SEME) and its possible impact on the outcomes of elections”, PNAS 2015, vol. 112 no. 33 Zemel, Wu, Swersky, Pitassi, and Dwork. “Learning Fair Representations.”, ICML (3), volume 28 of JMLR Proceedings, page 325-333 Chinese Social Credit System: volkskrant.nl/buitenland/china-rates-its-own-citizens-including-online- behaviour~a3979668/ bbc.com/news/world-asia-china-34592186 newyorker.com/news/daily-comment/how-china-wants-to-rate-its-citizens Predictive Policing: azavea.com/blogs/newsletter/v8i5/hunchlab-2-0-defining-the-future-of-predictive- policing/ popsci.com/want-to-prevent-another-ferguson predpol.com Political Manipulation: newstatesman.com/politics/2014/06/facebook-could-decide-election-without- anyone-ever-finding-out politico.com/magazine/story/2015/08/how-google-could-rig-the-2016- election-121548 Twins: wbay.com/2015/10/23/twins-denied-drivers-permit-because-dmv-cant-tell-them- apart/ ALPR: Greene v. San Francisco, 9th Cir. Court Image credit: Freddy Martinez