The social dilemma of autonomous vehicles · PDF fileThe social dilemma of autonomous vehicles...

14
The social dilemma of autonomous vehicles Jean-François Bonnefon 1 , Azim Shariff 2 , and Iyad Rahwan 3 1 Toulouse School of Economics, Institute for Advanced Study in Toulouse, Center for Research in Management, University of Toulouse, France 2 Department of Psychology, University of Oregon, Eugene, OR 97403, USA 3 The Media Lab, Massachusetts Institute of Technology, Cambridge, MA 02139, USA Autonomous Vehicles (AVs) should reduce traffic accidents, but they will sometimes have to choose between two evils—for example, running over pedestrians or sacrificing itself and its passenger to save them. Defining the algorithms that will help AVs make these moral decisions is a formidable challenge. We found that participants to six MTurk studies approved of utilitarian AVs (that sacrifice their passengers for the greater good), and would like others to buy them; but they would themselves prefer to ride in AVs that protect their passengers at all costs. They would disapprove of enforcing utilitarian AVs, and would be less willing to buy such a regulated AV. Accordingly, regulating for utilitarian algorithms may paradoxically increase casualties by postponing the adoption of a safer technology. Correspondence should be addressed to [email protected] The year 2007 saw the completion of the first benchmark test for autonomous driving in realistic urban environments [1, 2]. Since then, Autonomous Vehicles (AVs) such as the Google Car covered thousands of miles of real-road driving [3]. AVs promise world- changing benefits, by increasing traffic efficiency [4], reducing pollution [5], and This is the author’s version of the work. It is posted here by permission of the AAAS for personal use, not for redistribution. The definitive version was published in Science on 24 Jun 2016: Vol. 35, DOI: 10.1126/science.aaf2654. Supplementary Materials and data are available on the Science article page.

Transcript of The social dilemma of autonomous vehicles · PDF fileThe social dilemma of autonomous vehicles...

The social dilemma of autonomous vehicles

Jean-François Bonnefon1, Azim Shariff

2, and Iyad Rahwan

∗3

1 Toulouse School of Economics, Institute for Advanced Study in Toulouse, Center for

Research in Management, University of Toulouse, France

2 Department of Psychology, University of Oregon, Eugene, OR 97403, USA

3 The Media Lab, Massachusetts Institute of Technology, Cambridge, MA 02139, USA

Autonomous Vehicles (AVs) should reduce traffic accidents, but they will sometimes

have to choose between two evils—for example, running over pedestrians or sacrificing

itself and its passenger to save them. Defining the algorithms that will help AVs make

these moral decisions is a formidable challenge. We found that participants to six MTurk

studies approved of utilitarian AVs (that sacrifice their passengers for the greater good),

and would like others to buy them; but they would themselves prefer to ride in AVs that

protect their passengers at all costs. They would disapprove of enforcing utilitarian AVs,

and would be less willing to buy such a regulated AV. Accordingly, regulating for

utilitarian algorithms may paradoxically increase casualties by postponing the adoption of

a safer technology.

∗Correspondence should be addressed to [email protected]

The year 2007 saw the completion of the first benchmark test for autonomous driving in

realistic urban environments [1, 2]. Since then, Autonomous Vehicles (AVs) such as the

Google Car covered thousands of miles of real-road driving [3]. AVs promise world-

changing benefits, by increasing traffic efficiency [4], reducing pollution [5], and

This is the author’s version of the work. It is posted here by permission of the AAAS for personal use, not for redistribution. The definitive version was published in Science on 24 Jun 2016: Vol. 35, DOI: 10.1126/science.aaf2654. Supplementary Materials and data are available on the Science article page.

eliminating up to 90% of traffic accidents [6]. Not all crashes will be avoided, though,

and some crashes will require AVs to make difficult ethical decisions, in cases which

involve unavoidable harm [7]. For example, the AV may avoid harming several

pedestrians by swerving and sacrificing a passer-by, or the AV may be faced with the

choice of sacrificing its own passenger to save one or more pedestrians (Fig. 1).

Although these scenarios appear unlikely, even low-probability events are bound to occur

with millions of AVs on the road. Moreover, even if these situations were never to arise,

programming must nevertheless include decision rules about what to do such in

hypothetical situations. Thus these types of decisions need be made well before AVs

become a global market. Distributing harm is a decision that is universally considered to

fall within the moral domain [8, 9]. Accordingly, the algorithms that control AVs will

need to embed moral principles guiding their decisions in situations of unavoidable harm

[10]. Manufacturers and regulators will need to accomplish three potentially incompatible

objectives: being consistent, not causing public outrage, and not discouraging buyers.

However, pursuing these objectives may lead to moral inconsistencies. Consider for

example the case displayed in Fig. 1a, and assume that the most common moral attitude

is that the AV should swerve. This would fit a utilitarian moral doctrine [11], according to

which the moral course of action is to minimize casualties. But consider then the case

displayed in Fig. 1c. The utilitarian course of action, in that situation, would be for the AV

to swerve and kill its passenger—but AVs programmed to follow this course of action

might discourage buyers who believe their own safety should trump other considerations.

Even though such situations may be exceedingly rare, their emotional saliency is likely to

give them broad public exposure and a disproportionate weight in individual and public

decisions about AVs. To align moral algorithms with human values, we must start a

collective discussion about the ethics of AVs, that is, the moral algorithms that we are

willing to accept as citizens and to be subjected to as car owners. Here we initiate the

data-driven study of driverless car ethics, inspired by the methods of experimental ethics

[12].

We conducted six online surveys between June and November 2015. All studies were

programmed on Qualtrics survey software and recruited participants (USA residents only)

from the Mechanical Turk platform, for a compensation of 25 cents. The experimental

ethics literature largely relies on MTurk respondents, with robust results, even though

Turk respondents are not necessarily representative of the US population [13,14]. A

possible concern with MTurk studies is that some participants may already be familiar

with testing materials, when these materials are used by many research groups; but this

concern does not apply to our testing materials, which have never been used in a

published MTurk study so far.

In all studies, participants provided basic demographic information. Regression analyses

(see Supplementary online materials) showed that enthusiasm for self-driving cars was

consistently greater for younger, male participants. Accordingly, all subsequent analyses

included age and sex as covariates. The last item in every study was an easy question

relative to the traffic situation that participants had just considered, e.g., how many

pedestrians were on the road. Participants who failed this attention check (typically 10%

of the sample) were discarded from subsequent analyses.

Detailed statistical results for all studies are provided in Supplementary online materials

(Tables S1-S8). Overall, participants strongly agreed that it would be more moral for AVs

to sacrifice their own passengers, when this sacrifice would save a greater number of lives

overall.

In Study 1 (n = 182), 76% of participants thought that it would be more moral for AVs to

sacrifice one passenger, rather than kill ten pedestrians (with a 95% confidence interval of

69—82). These same participants were later asked to rate which was the most moral way

to program AVs, on a scale from 0 (protect the passenger at all costs) to 100 (minimize

the number of casualties). They overwhelmingly expressed a moral preference for

utilitarian AVs programmed to minimize the number of casualties (median = 85, Fig. 2a).

Participants were not quite so sure that AVs would be programmed that way though (67%

thought so, with a median rating of 70). Thus, participants were not worried about AVs

being too utilitarian, as often portrayed in science-fiction. If anything, they imagined

future AVs as being less utilitarian than they should be.

In Study 2 (n = 451), participants were presented with dilemmas that varied the number

of pedestrians’ lives that could be saved, from 1 to 100. Participants did not think that

AVs should sacrifice their passenger when only one pedestrian could be saved (with an

average approval rate of 23%), but their moral approval increased with the number of

lives that could be saved (p < 0.001) up to approval rates consistent with the 76%

observed in Study 1 (Fig. 2b),.

Participants’ approval of passenger sacrifice was even robust to treatments in which they

had to imagine themselves and another person in the AV, a family member in particular

(Study 3, n = 259). Imagining that a family member was in the AV negatively impacted

the morality of the sacrifice, as compared to imagining oneself alone in the AV (p =

0.003), but even in that strongly aversive situation, the morality of the sacrifice was still

rated above the midpoint of the scale, with a 95% confidence interval of 54–66 (Fig. 3a).

But this study presents the first hint of a social dilemma. On a scale of 1–100,

respondents were asked to indicate how likely they would be to buy an AV programmed

to minimize casualties (which would, in these circumstances, sacrifice them and their co-

rider family member), and how likely they would be to buy an AV programmed to

prioritize protecting its passengers, even if it meant killing 10 or 20 pedestrians. Al-

though the reported likelihood of buying an AV was low even for the self-protective

option (median = 50), respondents indicated a significantly lower likelihood (p < .001) of

buying the AV when they imagined the situation in which they and their family member

would be sacrificed for the greater good (median = 19). In other words, even though

participants still agreed that utilitarian AVs were the most moral, they preferred the self-

protective model for themselves.

Study 4 (n = 267) offers another demonstration of this phenomenon. Participants were

given 100 points to allocate between different types of algorithms, once to indicate how

moral the algorithms were, once to indicate how comfortable they were for other AVs to

be programmed that way, and once for how likely they would be to buy an AV

programmed that way. For one of the algorithms, the AV would always swerve when it

was about to run over people on the road. Fig 3b shows the points allocated to the AV

equipped with this algorithm, in three situations: when it swerved into a pedestrian to

save 10 people, when it killed its own passenger to save 10 people, and when it swerved

into a pedestrian to save just one other pedestrian. The algorithm that swerved into one to

save 10 always received many points, and the algorithm that swerved into one to save one

always received few points. The algorithm that would kill its passenger to save 10

presented a hybrid profile. Just as the high-valued algorithm, it received high marks for

morality (median budget share = 50), and was considered a good algorithm for other

people to have (median budget share = 50). But in terms of purchase intention, it received

significantly fewer points than the high-valued algorithm (p < 0.001), and was in fact

closer to the low-valued algorithms (median budget share = 33). Once more, it appears

that people praise utilitarian, self-sacrificing AVs, and welcome them on the road, without

actually wanting to buy one for themselves.

This is the classic signature of a social dilemma, in which everyone has a temptation to

free-ride instead of adopting the behavior that would lead to the best global outcome. One

typical solution in this case is for regulators to enforce the behavior leading to the best

global outcome; indeed, there are many other examples of tradeoff by people and

governments that involve harm [15-17]. For example, some citizens object to regulations

that require to immunize children before they start school, minimizing the perceived risk

of harm to their child, while increasing the risk to others; and recognition of the threats of

environmental degradation have prompted government regulations aimed at curtailing

harmful behaviors for the greater good. But would people approve of government

regulations imposing utilitarian algorithms in AVs, and would they be more likely to buy

AVs under such a regulation?

In Study 5 (n = 376), we asked participants about their attitudes towards legally enforcing

utilitarian sacrifices. Participants considered scenarios in which either a human driver or a

control algorithm had an opportunity to self-sacrifice in order to save 1 or 10 pedestrians

(Fig 3c). As usual, the perceived morality of the sacrifice was high, and about the same

whether the sacrifice was performed by a human or by an algorithm (median = 70). When

asked whether they would agree to see such moral sacrifices legally enforced, their

agreement was higher for algorithms than for human drivers (p < 0.002), but the average

agreement still remained below the midpoint of the 0–100 scale in each scenario.

Agreement was highest in the scenario where algorithms saved 10 lives, with a 95%

confidence interval of 33–46.

Finally, in Study 6 (n = 393), we asked participants specifically about their likelihood of

purchasing the AVs whose algorithms had been regulated by the government. Participants

were presented with scenarios in which they were either riding alone, with an unspecified

family member, or with their child. As in the previous studies the scenarios depicted a

situation in which the algorithm that controlled the AV could sacrifice its passengers to

minimize casualties on the road. Participants indicated whether it was the duty of the

government to take regulations that would minimize the casualties in such circumstances,

whether they would consider the purchase of an AV under such a regulation, and whether

they would consider purchasing an AV under no such regulation. As shown in Fig 3d,

people were reluctant to accept governmental regulation of utilitarian AVs. Even in the

most favorable condition, when they were only imagining themselves being sacrificed to

save 10 pedestrians, the 95% confidence interval for whether people thought it was

appropriate for the government to regulate this sacrifice was only 36–48. Finally and

importantly, participants were vastly less likely to consider purchasing an AV with such

regulation than without (p < 0.001). The median expressed likelihood of purchasing an

unregulated AV was 59, compared to 21 for purchasing a regulated AV. This is a huge gap

from a statistical perspective, but it must be understood as reflecting the state of public

sentiment at the very beginning of a new public issue, and thus not guaranteed to persist.

Three groups may be able to decide how AVs handle ethical dilemmas: the consumers

who buy the AVs; the manufacturers that must program the AVs in the first place; and the

government which may regulate the kind of programming that manufacturers can offer

and consumers can choose from. While manufacturers may engage in advertising and

lobbying to influence the preferences of consumers and the regulations that the

governments may pass, a critical collective problem consists of deciding whether

governments should regulate the moral algorithms that manufacturers offer to consumers.

Our findings suggests that regulation may be necessary, but at the same time counter-

productive. Moral algorithms for AVs create a social dilemma [18, 19]. Even though

people seem to agree that everyone would be better off if AVs were utilitarian (in the

sense of minimizing the number of casualties on the road), they all have a personal

incentive to ride in AVs that will protect them at all costs. Accordingly, if both self-

protective and utilitarian AVs were allowed on the market, few would be willing to ride in

utilitarian AVs, even though they would prefer others to do so. Regulation may provide a

solution to this problem, but it faces two difficulties. First, most people seem to

disapprove of a regulation that would enforce utilitarian AVs. Second and worse, our

results suggest that such a regulation could significantly delay the adoption of AVs—

which means that the lives saved by making AVs utilitarian may be outnumbered by the

deaths caused by delaying the adoption of the AVs altogether, with their other safety

enhancing technologies. This is a challenge that should be on the mind of carmakers and

regulators alike.

Moral algorithms for AVs will need to tackle more intricate decisions than that we have

considered in our surveys. For example, our scenarios did not feature any uncertainty

about decision outcomes, but a collective discussion about moral algorithms will have to

tackle the concepts of expected risk, expected value, and blame assignment. Is it

acceptable for an AV to avoid a motorcycle by swerving into a wall, considering that the

probability of survival is greater for the passenger of the AV, than for the rider of the

motorcycle? Should AVs take the ages of the passengers and pedestrians into account

[20]? If a manufacturer offers different versions of its moral algorithm, and a buyer

knowingly chose one of them, is the buyer to blame for the harmful consequences of the

algorithm’s decisions? Such liability considerations will need to accompany existing

discussions of regulation [21], and we hope that psychological studies inspired by our

own will be able to inform this discussion.

Figuring out how to build ethical autonomous machines is one of the thorniest challenges

in artificial intelligence today [22]. As we are about to endow millions of vehicles with

autonomy, taking algorithmic morality seriously has never been more urgent. Our data-

driven approach highlights how the field of experimental ethics can give us key insights

into the moral, cultural and legal standards that people expect from autonomous driving

algorithms. For the time being, there seems to be no easy way to design algorithms that

would reconcile moral values and personal self-interest, let alone across different cultures

with different moral attitudes to life-life tradeoffs [23]—but public opinion and social

pressure may very well shift as this conversation progresses.

References

[1]  Montemerlo B et al. Junior: The Stanford entry in the urban challenge. Journal of

Field Robotics, 25:569–597, 2008.

[2]  Urmson C et al. Autonomous driving in urban environments: Boss and the urban challenge. Journal of Field Robotics, 25:425–266, 2008.

[3]  Waldrop MM. Autonomous vehicles: No drivers required. Nature, 518:20–23, 2015.

[4]  Van Arem B, Van Driel CJ, Visser R. The impact of cooperative adaptive cruise

control on traffic-flow characteristics. IEEE Transactions on Intelligent Transportation

Systems, 7:429–436, 2006.

[5]  Spieser K et al. Toward a systematic approach to the design and evaluation of

automated mobility-on-demand systems: A case study in Singapore. In Meyer G, Beiker

S, editor, Road Vehicle Automation, pages 229–245. Springer, 2014.

[6]  Gao P, Hensley R, Zielke A. A roadmap to the future for the auto industry, 2014.

[7]  Goodall NJ. Machine ethics and automated vehicles. In Meyer G, Beiker S, editor,

Road Vehicle Automation, pages 93–102. Springer, 2014.

[8]  Gray K, Waytz A, and Young L. The moral dyad: A fundamental template unifying

moral judgment. Psychological Inquiry, 23:206–215, 2012.

[9]  Haidt J. The righteous mind: Why good people are divided by politics and religion.

Pantheon Books, 2012.

[10]  Wallach W, Allen C. Moral machines: Teaching robots right from wrong. Oxford

University Press, 2008.

[11]  Rosen F. Classical utilitarianism from Hume to Mill. Routledge, 2005.

[12]  Greene JD. Moral tribes: Emotion, reason and the gap between us and them. Atlantic Books, 2014.

[13] Côté S, Piff PK, Willer R. For whom do the ends justify the means? Social class and

utilitarian moral judgment. Journal of Personality and Social Psychology, 104: 490-503.,

2013

[14] Everett JAC, Pizarro DA, Crockett MJ. Inference of trustworthiness from intuitive

moral judgments. Journal of Experimental Psychology: General, 2016.

[15] Kass NE. An ethics framework for public health. American Journal of Public

Health, 91: 1776–1782, 2001.

[16] Sunstein CR, Vermeule A. Is capital punishment morally required? Acts omissions,

and life-life tradeoffs. Stanford Law Review, 58: 703–750, 2005.

[17] Dietz T, Ostrom E, Stern PC. The struggle to govern the commons. Science, 302:

1907-1912, 2003.

[18]  Dawes RM. Social dilemmas. Annual Review of Psycholology, 31:169–193, 1980.

[19]  Van Lange PAM, Joireman J, Parks CD, Craig D, Van Dijk E. The psychology of

social dilemmas: A review. Organizational Behavior and Human Decision Processes,

120:125–141, 2013.

[20] Posner EA, Sunstein CR. Dollars and death. The University of Chicago Law Review,

72: 537–598, 2005.

[21] Vladeck DC. Machines without principals: Liability rules and Artificial Intelligence.

Washington Law Review, 89: 117-150, 2014

[21]  Deng B. Machine ethics: The robot’s dilemma. Nature, 523:24–26, 2015.

[22] Gold N, Colman AM, Pulford, BD. Cultural differences in response to real-life and

hypothetical trolley problems. Judgment and Decision Making, 9: 65–76, 2014.

Bonnefon gratefully acknowledges support through the ANR-Labex IAST. This research

was supported by internal funds from the University of Oregon to Shariff. Rahwan is

grateful for financial support from Reid Hoffman. Data files have been uploaded as

Supplementary Online Material.

Figure 1: Three traffic situations involving imminent unavoidable harm. The car must

decide between (a) killing several pedestrians or one passer by, (b) killing one pedestrian or killing its own passenger, (c) killing several pedestrians or killing its own passenger.

Figure 2: Considering the greater good vs the life of the passenger. In studies 1 and 2,

when asked which would be the most moral way to program AVs, participants expressed

a preference for AVs programmed to kill their passenger for the greater good. This

preference was strong as soon as 5 lives or more could be saved (left panel shows

detailed results for 10 lives). On average, participants were more confident that AVs

should pursue the greater good than they were confident that AVs would be programmed

to do so. Boxes show the 95% confidence interval of the mean.

Fig. 3: Towards regulation and purchase (studies 3-6). Boxes show the 95% confidence

interval of the mean. In all studies, participants expressed a moral preference for AVs

sacrificing their passengers to save a greater number of pedestrians. This moral

preference was robust to situations in which participants imagined themselves in the AV

in the company of a coworker, a family member, or their own child. However, (a and b)

participants did not ex- press a comparable preference for buying utilitarian AVs,

especially when they thought of family members riding in the car, and (c and d) they

disapproved of regulations enforcing utilitarian algorithms for AVs, and indicated that

they would be less likely to purchase an AV under such a regulation.

a b cSWERVESWERVE STAYSTAYSWERVESWERVE STAYSTAYSWERVESWERVE STAYSTAYSWERVESWERVE STAYSTAYSWERVESWERVE STAYSTAY

0

20

40

60

0

20

40

60

What w

ill AVs do?W

hat should AVs do?

0 25 50 75 100From 0 = 'Protect driver at all costs'

to 100 = 'Minimize casualties on the road'

Num

ber o

f Res

pons

es

Study 1A

0.25

0.50

0.75

1 2 5 20 100Number of lives saved by sacrifice

Prop

ortio

n w

ho a

gree

Will AVs sacrifice? Should AVs sacrifice?

Study 2BStudy 1 Study 2

20

40

60

80

Moralityof sacrifice

BuyMinimize

BuyProtective

Agre

emen

t with

Mor

ality

(or)

Willi

ngne

ss to

Buy

Alone with Coworker with Family

Study 3a

20

40

60

80

Moralityof Sacrifice

Wants thisfor other cars

Wants thisfor my car

Budg

et S

hare

Kill Pedestrianto Save Ten

Kill Passengerto Save Ten

Kill Pedestrianto Save One

Study 4b

20

40

60

80

Moralityof Sacrifice

LegallyEnforce Sacrifice

Agre

emen

t with

Mor

ality

(or)

Agre

emen

t with

Leg

al E

nfor

cem

ent

HumanSaves One

AlgorithmSaves One

HumanSaves Ten

AlgorithmSaves Ten

Study 5c

20

40

60

80

Moralityof Sacrifice

RegulatedSacrifice

Purchase ifregulated

Purchase ifnot regulated

Agre

emen

t with

Mor

ality

and

Reg

ulat

ion

(or)

Willi

ngne

ss to

Pur

chas

eAlone with Family with own Child

Study 6d