2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

13
2 Dec. 2017, JADS Discussion on Fairness in Data Science Moderated by Dr. Maurits C. Kaptein

Transcript of 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Page 1: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

2 Dec. 2017, JADS

Discussion on Fairness in Data Science

Moderated by Dr. Maurits C. Kaptein

Page 2: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fairness: How to avoid unfair conclusions?

Accuracy: How do we guarantee accuracy?

Confidentiality: How to not reveal secrets?

Transparency: How to make answers indisputable?

Page 3: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

First focus on Fairness; lets assume the model output is

accurate and transparent…

Page 4: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

What is Fairness?

- Non discriminating?

- Broader definitions?

- Veil of ignorance?

Canonical example of unfair Data Science: Higher bail-out charges for African Americans based on predicted risk of recurrence

Search for examples….

Page 5: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fair? (show of hands)

Decision to search a person in the airport based on

membership of a terrorist group.

Page 6: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fair?

Decision to search a person in the airport based on

religious beliefs.

Page 7: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fair?

Decision to search a person in the airport based on a set of

features that does not include religious belief, but does

predict it accurately.

Page 8: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fair?

Decision to search a person in the airport based on race.

Page 9: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fairness implemented:

- Is a distinct set of “non-permitted” features a sufficient

approach to ensure fairness? (again, assume the models

are accurate, etc.)

Page 10: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fairness implemented:

- Can we identify a such a feature set?

- How do we select these features? (legal framework, …)

- What if the non-permitted features are directly related to

permitted features?

Page 11: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fair?

Decision to change a medical treatment – to the benefit of

the patient – based on race.

Page 12: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Fairness implemented:

- Can our permitted feature set be context independent?

Or does each problem need its own list? - Societal vs. personal gains?

Page 13: 2. Workshop Responsible Data Science - Data science without prejudice by Maurits Kaptein

Discussion points:

- Can fairness be formalized?

- Is the current legal framework sufficient?

- Can we enforce / regulate fairness?

- …?