discrimination in decision making humans vs machines
play

Discrimination in Decision Making: Humans vs. Machines Muhammad - PowerPoint PPT Presentation

Discrimination in Decision Making: Humans vs. Machines Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, Krishna P. Gummadi Max Planck Institute for Software Systems Machine decision making q Refers to data-driven algorithmic


  1. Discrimination in Decision Making: Humans vs. Machines Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez-Rodriguez, Krishna P. Gummadi Max Planck Institute for Software Systems

  2. Machine decision making q Refers to data-driven algorithmic decision making q By learning over data about past decisions q To assist or replace human decision making q Increasingly being used in several domains q Recruiting: Screening job applications q Banking: Credit ratings / loan approvals q Judiciary: Recidivism risk assessments q Journalism: News recommender systems

  3. The concept of discrimination q Discrimination is a special type of unfairness q Well-studied in social sciences q Political science q Moral philosophy q Economics q Law q Majority of countries have anti-discrimination laws q Discrimination recognized in several international human rights laws q But, less-studied from a computational perspective

  4. Why, a computational perspective? 1. Datamining is increasingly being used to detect discrimination in human decision making q Examples: NYPD stop and frisk, Airbnb rentals

  5. Why, a computational perspective? 2. Learning to avoid discrimination in data-driven (algorithmic) decision making q Aren’t algorithmic decisions inherently objective? q In contrast to subjective human decisions q Doesn’t that make them fair & non-discriminatory? q Objective decisions can be unfair & discriminatory!

  6. Why, a computational perspective? q Learning to avoid discrimination in data-driven (algorithmic) decision making q A priori discrimination in biased training data q Algorithms will objectively learn the biases q Learning objectives target decision accuracy over all users q Ignoring outcome disparity for different sub-groups of users

  7. Our agenda: Two high-level questions How to detect discrimination in decision making? 1. Independently of who makes the decisions q Humans or machines q How to avoid discrimination when learning? 2. q Can we make algorithmic decisions more fair? q If so, algorithms could eliminate biases in human decisions q Controlling algorithms may be easier than retraining people

  8. This talk How to detect discrimination in decision making? 1. Independently of who makes the decisions q Humans or machines q How to avoid discrimination when learning? 2. q Can we make algorithmic decisions more fair? q If so, algorithms could eliminate biases in human decisions q Controlling algorithms may be easier than retraining people

  9. The concept of discrimination q A first approximate normative / moralized definition: wrongfully impose a relative disadvantage on persons based on their membership in some salient social group e.g., race or gender

  10. The concept of discrimination q A first approximate normative / moralized definition: wrongfully impose a relative disadvantage on persons based on their membership in some salient social group e.g., race or gender

  11. The devil is in the details q What constitutes a salient social group? q A question for political and social scientists q What constitutes relative disadvantage? q A question for economists and lawyers q What constitutes a wrongful decision? q A question for moral-philosophers q What constitutes based on? q A question for computer scientists

  12. Discrimination: A computational perspective q Consider binary classification using user attributes A 1 A 2 … A m Decision User 1 x 1,1 x 1,2 … x 1,m Accept x 2,1 x 2,m Reject User 2 User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept

  13. Discrimination: A computational perspective q Consider binary classification using user attributes SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Some attributes are sensitive, others non-sensitive

  14. Discrimination: A computational perspective q Consider binary classification using user attributes SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Some attributes are sensitive, others non-sensitive Decisions should not be based on sensitive attributes!

  15. What constitutes “not based on”? q Most intuitive notion: Ignore sensitive attributes q Fairness through blindness or veil of ignorance q When learning, strip sensitive attributes from inputs q Avoids disparate treatment q Same treatment for users with same non-sensitive attributes q Irrespective of their sensitive attribute values q Situational testing for discrimination discovery checks for this condition

  16. Two problems with the intuitive notion When users of different sensitive attribute groups have different non-sensitive feature distributions, we risk Disparate Mistreatment 1. Even when training data is unbiased, sensitive attribute groups q might have different misclassification rates Disparate Impact 2. When labels in training data are biased, sensitive attribute groups q might see different beneficial outcomes to different extents Training data bias due to past discrimination q

  17. Background: Two points about learning To learn, we define & optimize a risk (loss) function 1. q Over all examples in training data q Risk function captures inaccuracy in prediction q So learning is cast as an optimization problem For efficient learning (optimization) 2. q We define loss functions so that they are convex

  18. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept

  19. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Suppose users are of two types: blue and pink

  20. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q Blue users might have a different risk / loss than red users!

  21. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q Stripping sensitive attributes does not help!

  22. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q To avoid disp. mistreatment, we need L(W) = L(W)

  23. Origins of disparate mistreatment SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept q Minimizing L(W) , does not guarantee L(W) and L (W) are equally minimized q Put differently, we need:

  24. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Reject … … … … … User n x n,1 x n,2 x n,m Accept

  25. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels!

  26. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels! q Classifier will learn to make biased decisions q Using sensitive attributes (SAs)

  27. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels! q Stripping SAs does not fully address the bias

  28. Origins of disparate impact SA 1 NSA 2 … NSA m Decision User 1 x 1,1 x 1,2 … x 1,m Accept User 2 x 2,1 x 2,m Reject User 3 x 3,1 x 3,m Accept … … … … … User n x n,1 x n,2 x n,m Reject q Suppose training data has biased labels! q Stripping SAs does not fully address the bias q NSAs correlated with SAs will be given more / less weights q Learning tries to compensate for lost SAs

  29. Analogous to indirect discrimination q Observed in human decision making q Indirectly discriminate against specific user groups using their correlated non-sensitive attributes q E.g., voter-id laws being passed in US states q Notoriously hard to detect indirect discrimination q In decision making scenarios without ground truth

Recommend


More recommend