Aditya V. Nori Microsoft Research Cambridge Joint work with Samuel - - PowerPoint PPT Presentation

aditya v nori
SMART_READER_LITE
LIVE PREVIEW

Aditya V. Nori Microsoft Research Cambridge Joint work with Samuel - - PowerPoint PPT Presentation

Aditya V. Nori Microsoft Research Cambridge Joint work with Samuel Drews, Aws Albarghouthi, Loris DAntoni (University of Wisconsin-Madison) Data is everywhere! Data analysis is big part of todays software Increasingly developers


slide-1
SLIDE 1

Aditya V. Nori

Microsoft Research Cambridge

Joint work with Samuel Drews, Aws Albarghouthi, Loris D’Antoni (University of Wisconsin-Madison)

slide-2
SLIDE 2

Data is everywhere!

▪ Data analysis is big part of today’s software ▪ Increasingly developers creating and using machine learning models ▪ Increasingly developers working with data that is incomplete, inaccurate, approximate

2

slide-3
SLIDE 3

▪ Brin

ingi ging ng da data ta in into pr progr grams ams

  • Numerous new sources
  • Conversion

▪ Rea

easoni soning ng wit ith da data ta

  • What does correctness mean?

Reasoning Data sources

Cleaning, Combining, Statistics, Machine Learning

Action

Visualize, Report, Recommend 3

slide-4
SLIDE 4
slide-5
SLIDE 5
slide-6
SLIDE 6
slide-7
SLIDE 7
slide-8
SLIDE 8

How do we prove that a program does not discriminate?

slide-9
SLIDE 9

 Theoreticians

▪ How do we formalise fairness?

 Machine learning researchers

▪ How do we learn fair models?

 Security/privacy researchers

▪ How do we detect bias in black-box algorithms

 Legal scholars

▪ How do we regulate algorithmic decision making?

slide-10
SLIDE 10

 EU GDPR

R (2018) 018)

▪ “data subject’s explicit consent” ▪ “right to explanation”  Whit

ite e House se repo port t (2014) 14)

▪ “Powerful algorithms … raise the potential of encoding

  • ding discri

scrimin minat ation ion in automat tomated ed decisi cisions.”

 White House report (recently …) ▪ “Federal agencies that use AI-based systems to make or provide decision

support for consequential decisions about individuals should take extra care to ensure e the efficac icacy y and fairn irness ess of those systems, based on evidenc dence- based ed ve verifi ificat cation

  • n and validat

idation ion.”

slide-11
SLIDE 11

1)

Fairness as a program property

2)

Automatic proofs of (un)fairness

slide-12
SLIDE 12
slide-13
SLIDE 13
slide-14
SLIDE 14

population model Hoare triple ☺

slide-15
SLIDE 15
slide-16
SLIDE 16
slide-17
SLIDE 17

by definition of conditional probability

slide-18
SLIDE 18

 𝛲: set of all possible execution paths in dec(popModel())  𝑞(𝜌): probability that 𝜌 ∈ 𝛲

slide-19
SLIDE 19

 𝛲: set of all possible execution paths in dec(popModel())  𝑞(𝜌): probability that 𝜌 ∈ 𝛲

slide-20
SLIDE 20

 𝛲: set of all possible execution paths in dec(popModel())  𝑞(𝜌): probability that 𝜌 ∈ 𝛲

What does this mean?

slide-21
SLIDE 21
slide-22
SLIDE 22

Each path is uniqu iquel ely represent esented ed by 3 real values

Ide dea

represent paths 𝛲ℎ𝑛 as a region 𝜒 ⊆ ℝ3

and compute:

slide-23
SLIDE 23

 Volume: ׬

𝜒 1 𝑒𝑓 𝑒𝑞 𝑒𝑧

𝜒 𝑞𝑧(𝑧)

slide-24
SLIDE 24
slide-25
SLIDE 25
slide-26
SLIDE 26
slide-27
SLIDE 27
slide-28
SLIDE 28

 Represent all executions as an SMT formula 𝜒  Compute the weighted volume of 𝜒

Volume of a polytope is #P-hard [Dyer and Frieze, 1988]

slide-29
SLIDE 29
slide-30
SLIDE 30

 Rectangles are easy!

slide-31
SLIDE 31

 There are inf

nfin init itely ely man any y rec ecta tang ngles les

slide-32
SLIDE 32

 Hyperrectangular decomposition ▪ consider all hyperrectangles in 𝜒  Hyperrectangular sampling ▪ Iteratively sample 𝐼 ⇒ 𝜒

slide-33
SLIDE 33
slide-34
SLIDE 34
slide-35
SLIDE 35
slide-36
SLIDE 36
slide-37
SLIDE 37
slide-38
SLIDE 38

 Ide

deal al so soluti ution:

  • n: sample with the following objective
slide-39
SLIDE 39

 Approximate densities with step functions ▪ area under a step function is a linear formula

slide-40
SLIDE 40

 Maintains a lower-bound on volume  Converges to the actual volume in the limit  Works for real closed fields  To compute upper-bound, negate formula

slide-41
SLIDE 41
slide-42
SLIDE 42
slide-43
SLIDE 43
slide-44
SLIDE 44

 Automatic proofs of (un)fairness for decision making

programs

 Future directions ▪ Scalability – application to real-world programs ▪ Explaining unfairness ▪ Repairing unfair programs

Fai airSquare: Pr Probabilistic Ver erifi fication for

  • r Pr

Program Fai airness Aws Albarghouthi, Loris D'Antoni, Samuel Drews, Aditya Nori, OOPSLA '17