iBet uBet web content aggregator. Adding the entire web to your favor.
iBet uBet web content aggregator. Adding the entire web to your favor.



Link to original content: https://www.academia.edu/124918728/Drivers_are_blamed_more_than_their_automated_cars_when_both_make_mistakes
(PDF) Drivers are blamed more than their automated cars when both make mistakes | Sohan Dsouza - Academia.edu
Academia.eduAcademia.edu

Drivers are blamed more than their automated cars when both make mistakes

2019, Nature Human Behaviour

ORE Open Research Exeter TITLE Drivers are blamed more than their automated cars when both make mistakes AUTHORS Awad, E; Levine, S; Kleiman-Weiner, M; et al. JOURNAL Nature Human Behaviour DEPOSITED IN ORE 14 October 2019 This version available at http://hdl.handle.net/10871/39189 COPYRIGHT AND REUSE Open Research Exeter makes this work available in accordance with publisher policies. A NOTE ON VERSIONS The version presented here may differ from the published version. If citing, you are advised to consult the published version for pagination, volume/issue and date of publication Drivers are blamed more than their automated cars when both make mistakes Edmond Awada,b,+ , Sydney Levinea,c,d,+ , Max Kleiman-Weinerc,d , Sohan Dsouzaa , Joshua B. Tenenbaumc,* , Azim Shariffe,* , Jean-François Bonnefona,f,* , and Iyad Rahwana,g,h,* a Media Lab, Massachusetts Institute of Technology, Cambridge MA, USA Department of Economics, University of Exeter Business School, Exeter, UK c Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, Cambridge MA, USA d Department of Psychology, Harvard University, Cambridge, MA, USA e Department of Psychology, University of British Columbia Vancouver, Canada f Toulouse School of Economics (TSM-Research), Centre National de la Recherche Scientifique, University of Toulouse Capitole, Toulouse, France. g Institute for Data, Systems and Society (IDSS), Massachusetts Institute of Technology, Cambridge MA, USA h Centre for Humans & Machines, Max-Planck Institute, Berlin, Germany + Joint first author * To whom correspondence should be addressed. e-mail: jbt@mit.edu; shariffa@uci.edu; jean-francois.bonnefon@tse-fr.eu; irahwan@mit.edu b 1 Abstract When an automated car harms someone, who is blamed by those who hear about it? Here, we asked human participants to consider hypothetical cases in which a pedestrian was killed by a car operated under shared control of a primary and a secondary driver, and to indicate how blame should be allocated. We find that when only one driver makes an error, that driver is blamed more, regardless of whether that driver is a machine or a human. However, when both drivers make errors in cases of human-machine shared-control vehicles, the blame attributed to the machine is reduced. This finding portends a public under-reaction to the malfunctioning AI components of automated cars and therefore has a direct policy implication: allowing the de-facto standards for shared-control vehicles to be established in courts by the jury system could fail to properly regulate the safety of those vehicles; instead, a top-down scheme (through federal laws) may be called for. Introduction Every year, about 1.25 million people die worldwide in car crashes [1]. Laws concerning principles of negligence currently adjudicate how responsibility and blame get assigned to the individuals who injure others in these harmful crashes. The impending transition to fully automated cars promises a radical shift in how blame and responsibility will get attributed in the cases where crashes do occur, but most agree that little or no blame will be attributed to the occupants in the car, who will, by then, be entirely removed from the decision-making loop [2]. However, before this era of fully automated cars arrives, we are entering a delicate era of shared control between humans and machines. This new moment signals a departure from our current system – where individuals have full control over their vehicles and thereby bear full responsibility for crashes (absent mitigating circumstances) – to a new system where blame and responsibility may be shared between a human and a machine driver. The spontaneous reactions people have to crashes that occur when a human and machine share control of a vehicle has at least two direct industry-shaping implications. First, at present, little is known about how the public is likely to respond to crashes that involve both human and machine drivers. This uncertainty has concrete implications: manufacturers price products to reflect the liability they expect to incur from the sale of those products. If manufacturers cannot assess the scope of the liability they will incur from automated vehicles (AVs), that uncertainty will translate to substantially inflated prices of AVs [2]. Moreover, the rate of the adoption of automated vehicles will be proportional to the cost to consumers to adopt the new technology [2].(The rate of the adoption of this technology is contingent on many other factors, including consumers’ understanding of the relative risks and benefits of using the cars. We do not mean to state that uncertainty about the scope of liability for manufacturers is the only factor impacting adoption, just that it is a significant one.) Accordingly, the uncertainty about the extent of corporate liability for automated vehicle crashes may be slowing down AV adoption [2], while people continue to die in car crashes each year. Clarifying how and when responsibility will be attributed to manufacturers in automated car crashes will be a first step in reducing this uncertainty and speeding the adoption of automated and eventually fully automated vehicles. The second direct implication of this work will be to forecast how a tort-based regulatory scheme (which is decided on the basis on jury decisions) is likely to turn out. Put another way, understanding how the public is likely to react to crashes that involve both a human and a machine driver will give us a hint to what standards will be established if we let jury decisions shape them. If our work uncovers systematic biases that are likely to impact juries and would impede the adoption 2 of automated cars, then it may make sense for federal regulations be put in place, which would preempt the tort system from being the avenue for establishing standards for these cars. Already, automated vehicle crashes are in the public eye. In May 2016, the first deadly crash of a Tesla Autopilot car occurred and the occupant of the car was killed. In a news release, Tesla explained: “Neither Autopilot nor the driver noticed the white side of the tractor-trailer against a brightly lit sky, so the brake was not applied” [3]. In March, 2018, the first automated car crash that killed a pedestrian occurred. A pedestrian that was crossing the street went unnoticed by both the car and the back-up driver. A few seconds before the crash, the car finally identified that it should be breaking but failed to do so. The driver also braked too late to avoid the collision. In both the Tesla and Uber fatal crashes, both the machine driver and the human driver should have taken action and neither driver did. The mistakes of both the machine and the human led to the crash. The National Highway Safety Traffic Administration (NHSTA) did an investigation of the Tesla incident and did not find Tesla at fault in the crash [4]. Likewise, Uber has been exonerated from criminal charges after an investigation by a county prosecutor [5]. Notably, press attention surrounding the Tesla incident was markedly skewed towards blaming the human driver for the crash, with rumors quickly circulating that the driver had been watching a Harry Potter movie [6], though upon further investigation it was discovered that there was no evidence grounding this claim [7]. Likewise, the fate of the Uber back-up driver remains unknown [5], with press attention focusing on the distracted nature of the driver [8]. The set of anecdotes around these two crashes begins to suggest a troubling pattern, namely, that humans might be blamed more than their machine partners in certain kinds of automated vehicle crashes. Was this pattern a fluke of the circumstances of the crash and the press environment? Or does it reflect something psychologically deeper that may color our responses to human-machine joint action, and in particular, when a human-machine pair jointly controls a vehicle? What we are currently witnessing is a gradual and multipronged increase toward full automation, going through several steps of shared control between user and vehicle, which may take decades due to technical and regulatory issues as well as attitudes of consumers towards adoption [9, 10] (see Figure 1). Some vehicles can take control over the actions of a human driver (e.g., Toyota’s ‘Guardian’) to perform emergency maneuvers. Other vehicles may do most of the driving, while requiring the user to constantly monitor the situation and be ready to take control (e.g., Tesla’s ‘Autopilot’). Unless clear or explicitly mentioned, we use “Human” and “User” interchangeably to refer to the person inside the car (being a driver or a passenger), and we use “Industry” and “Machine” interchangeably to refer to both company and car combined together. Our central question is this: when an automated car crashes and harms someone, how is blame and causal responsibility attributed to the human and machine drivers by people who hear about the crash? In this article, we use vignettes in which a pedestrian was hit and killed by a car being operated under shared control of a primary and a secondary driver and ask our participants to evaluate the crash on metrics of blame and causal responsibility. The cases we use are hypothetical (insofar as respondents know that they did not actually take place), but are not unrealistic, as they were designed to contain the relevant elements of events that could actually occur. We consider a large range of control regimes (see Figure 1), but the two main cases of interest are the instances of shared control where a human is the primary driver and the machine a secondary driver (“humanmachine”) and where the machine is the primary driver and the human the secondary driver (“machine-human”). We consider a simplified space of scenarios in which (a) the main driver makes the correct choice and the secondary driver incorrectly intervenes (“Bad Intervention”) and (b) the main driver makes an error and the secondary driver fails to intervene (“Missed Intervention”). Both scenarios end in a crash. For comparison, we also include analogous scenarios involving a single human driver (a regular car) or a single machine driver (a fully automated car) as well as two hypothetical two-driver cars (driven by two humans or two machines). We ask participants to 3 make evaluations of the human user and one representative of the machine, either the car itself, or the company that designed the car. Condition 1st 2nd Real World Implementation H — Regular H H — H M Guardian M H Autopilot M M — M — Fully Automated Bad Intervention Missed Intervention H Human Human Decision Omission Best outcome M Machine Machine Decision Commission Worst outcome Figure 1: Actions or action sequences for the different car types considered. Outline arrows indicate an action by a human, ‘H’, and solid arrows indicate an action by a machine, ‘M’. The top and bottom rows represent sole-driver cars, while all others represent dual-driver cars. A red arrow indicates a decision – whether action or inaction – that had the avoidable death of a pedestrian as the outcome. A blue arrow indicates a decision that does not result in any deaths. For example, the H+M type (real-world implementation is the Guardian system) has a human main driver and a machine standby driver. A Bad Intervention then involves the human staying on course (a non-lethal action, indicated by the outline of the straight, blue arrow) and the machine overriding that action, causing the death of the pedestrian (solid, angled, red arrow). A Missed Intervention involves the human staying on course to kill the pedestrian (outline, straight, red arrow) without intervention from the machine (solid, straight, red arrow). In Bad Intervention cases (see Methods – Case Description for details), the primary driver (be it human or machine) has made a correct decision to keep the car on course, which will avoid a pedestrian. Following this, the secondary driver makes the decision to swerve the car into the pedestrian. In these sorts of cases, we expect that the secondary driver (the only driver that makes a mistake) will be blamed more than the first driver. What is less clear is if people will assign blame and causal responsibility differently if this secondary driver is a human driver or a machine. Recent research suggests that humans may be blamed more than robots for making the same error 4 in the same situations [11]. In Missed Intervention cases, the primary driver has made an incorrect decision to keep the car on course (rather than swerving), which would cause the car to hit and kill a pedestrian. The secondary driver then neglects to swerve out of the way of the pedestrian. In these cases, the predictions for how participants will distribute blame and causal responsibility are less clear because both drivers make a mistake. As in the Bad Intervention cases, agent type (human or machine) may have an effect on blame and causal responsibility ratings. But unlike with Bad Intervention cases, Missed Intervention cases introduce the possibility that driver role (primary or secondary) may also impact judgments. It is possible that participants may shift responsibility and blame either toward the agent who contributed the most to the outcome (primary driver), or to the agent who had the last opportunity to act (secondary driver; [12, 13, 14, 15]). Under some regimes – such as Toyota’s Guardian – the user does most of the driving, but the decision to override (and thus to act last) pertains to the machine. Under others – such as Tesla’s Autopilot – the machine does most of the driving, but the decision to override pertains to the user. Results All studies used hypothetical vignettes that describe a crash (see Methods – Case Description for details on car regimes and intervention types, and see Supplementary Methods 1 for vignettes of studies 1-5). Study 1 Study 1 compared four kinds of cars with different regimes of control. Each car had a primary driver, whose job it was to drive the car, and a secondary driver, whose job it was to monitor the actions of the first driver and intervene when the first driver made an error. The car architectures of central interest were human primary-machine secondary (“human-machine”) and machine primaryhuman secondary (“machine-human”). We also included human-human and machine-machine architectures for comparison. This allowed us to see how blame was distributed in a dual-driver architecture when there was no difference in driver type (human or machine) in each of the driving roles (primary or secondary). Bad Interventions In Bad Intervention cases, two predictors were entered into a regression with blame and causal responsibility ratings as the outcome variable: (1) whether or not the driver made an error and (2) driver type (human or machine). The main finding is that whether or not the driver made an error was a significant predictor of ratings (see Table 1 – Column: Bad Intervention - Study 1). In other words, a driver that unnecessarily intervened, leading to the death of a pedestrian was blamed more than a driver that operated on the correct course – regardless of whether the driver was a human or machine. It is worth noting here that we did not detect a reliable effect of driver type (Human vs. Machine), once correcting for multiple comparisons (see Table 1 – Row: Human, Column: Bad Intervention - Study 1). We do not discuss this factor further in the Bad Intervention cases. Missed Interventions In Missed Intervention cases, blame and responsibility judgments cannot depend on whether or not a driver made an error because both drivers make errors in these cases. The main finding for these cases is that driver type – whether the driver is a human or machine – has a significant 5 a Treatment Bad Intervention Missed Intervention S2-M S1-MM S2-MH Agent critical cases S1-HM critical cases S1-MH Industry User S2-HM S1-HH S2-H 100 < first driver 0 last driver > 100 100 < first driver 0 last driver > 100 Blame and responsibility b Treatment Bad Intervention Missed Intervention S2-M S1-MM S1-MH Agent S1-HM Industry S2-MH User S2-HM S1-HH S2-H 100 < first driver 0 last driver > 100 100 < first driver 0 last driver > 100 Blame and responsibility Figure 2: Blame ratings for User and Industry in six car types. (a) Bar plot and (b) Dot plot. Data from Study 1 (S1: N = 786, Observations = 3, 144) and Study 2 (S2: N = 382, Observations = 1, 528). Ratings of blame and causal responsibility are aggregated (collectively referred to as blame, henceforth). Ratings of car and company are aggregated (collectively referred to as Industry, henceforth). The y-axis represents the six car types considered in Studies 1 and 2 (S1 and S2). Two car types, HM (human-machine) and MH (machine-human), were considered in both studies. The y-axis labels include the study and the car type. For example, S1-HM represents the Human-Machine regime ratings collected in Study 1. In the six car types, the xaxis labeling of first driver refers to the main driver, and the last driver refers to the secondary driver in dual-driver cars, and the sole driver in the sole-driver cars. For Bad Intervention, only one agent has erred (the last driver). This agent (whether User or Industry) is blamed more than the other agent (first driver; see Table 1). For Missed Intervention, in dual-driver cars (rows 2-7), both agents have erred. When human and machine are sharing control (inside the dotted rectangle), blame ratings of Industry drops significantly, regardless of the role of the machine (main or secondary). In Study 1, blame to Industry in S1-MH (m1 = 57.2) is significantly less than in S1-MM (m2 = 68), [t(760.6) = −5.05, p < .0001, m2 − m1 = 10.8, 95% CI for m2 − m1 is 6.6–15]. And blame to Industry in S1-HM (m1 = 53.4) is significantly less than in S1-MM (m2 = 68), [t(722.77) = −6.6042, p < .0001, m2 − m1 = 14.6, 95% CI for m2 − m1 is 10.2–19]. In Study 2, blame to Industry in S2-M (m1 = 75.6) is significantly more than in S2-MH (m2 = 59.5), [t(754.63) = −7.3885, p < .0001, m1 − m2 = 16.1, 95% CI for m1 − m2 is 11.8–20.3]; And is significantly more than in S2-HM (m3 = 48.51), [t(745.06) = 11.676, p < .0001, m1 − m3 = 27.1, 95% CI for m1 − m3 is 22.5–31.6]. 6 impact on ratings. Specifically, in these shared-control scenarios, where both human and machine have made errors, the machine driver is consistently blamed less than the human driver (Table 1 – Column: Missed Intervention - Study 1; Figure 2). The human-machine difference appears to be driven by a reduction in the blame attributed to machines when there is a human in the loop. This is evident when comparing both the humanmachine and machine-human instances of shared control to the machine-machine scenario. Note that the behaviors in these scenarios are identical, but how much a machine is blamed depends on whether it is sharing control with a human or operating both the primary and secondary driver role. When the machine is the primary driver, it is held significantly less blameworthy when its secondary driver is a human (m1 = 57.2) compared to when the secondary driver is also the machine (m2 = 68), t(760.6) = −5.0, p < .0001, difference in means: m2 − m1 = 10.8, 95% CI for m2 − m1 is 6.6–15. (All tests are two-tailed.) Similarly, when the machine is the secondary driver, it is held significantly less blameworthy when its primary driver is a human (m1 = 53.4), compared to when the primary driver is also the machine (m2 = 68), t(722.77) = −6.6, p < .0001, difference in means: m2 − m1 = 14.6, 95% CI for m2 − m1 is 10.2–19. Study 2 Study 2 compared the human-machine and machine-human shared control cars with two different baseline cars: a standard car, which is exclusively driven by a human, and a fully automated car, which is exclusively driven by a machine. This allowed us to both replicate the main results of Study 1 (the responses to Machine-Human and Human-Machine crashes) and also to see how blame was assigned differently to dual-driver cars as compared to sole-driver cars. The industry representative was varied (car and company) but this exploratory variable was not analyzed in this study and in subsequent studies. Bad Interventions We replicated the main results of Study 1. Namely, in Bad Intervention cases for the sharedcontrol cars (Machine-Human and Human-Machine), whether or not the driver made an error was a significant predictor of ratings (Table 1 – Column: Bad Intervention - Study 2; Figure 2). Missed Intervention We again replicated the main finding of Study 1. Driver type – whether the driver is a human or machine – has a significant impact on ratings. Specifically, in shared-control scenarios (MachineHuman and Human-Machine), where both human and machine have made errors, the machine driver is consistently blamed less than the human driver (Table 1 – Column: Missed Intervention - Study 2; Figure 2). As we noted in Study 1, the human-machine difference is driven by a reduction in the blame attributed to machines when there is a human in the loop. This is verified in Study 2 by comparing blame to the machine in the shared control cases with blame to the machine in the fully automated car (driven by a sole machine driver). In each case, blame to the machine in the shared control case is significantly lower than blame to the machine in the Fully Automated car: Fully Automated (m1 = 75.6) vs. Machine-Human (m2 = 59.5), t(754.63) = −7.4, p < .0001, m1 − m2 = 16.1, 95% CI for m1 − m2 is 11.8–20.3; vs. Human-Machine (m3 = 48.5), t(745.06) = 11.7, p < .0001, m1 − m3 = 27.1, 95% CI for m1 − m3 is 22.5–31.6. 7 Table 1: Regression analysis of data collected in studies 1-5 in the cases of Bad Intervention and Missed Intervention. Data from Studies 2 and 3 are limited to shared-control regimes in the table. “Human” refers to the type of agent in question (that is, human as compared to the baseline, machine), “Mistake” refers to whether the decision was a mistake (that is, the decision would have resulted in losing a life, or losing more lives in study 3), and “Last Driver” refers to the driver role (that is, the driver assumes the secondary role). All models include participant random effects and question (blame or causal responsibility) random effects, where applicable. Data were assumed to meet the requirements of the model. Blame and Causal Responsibility Bad Intervention Missed Intervention Study 1 Study 2 Study 3 Study 1 Study 2 Study 3 Study 4 Study 5 Human 2.141 (1.061) p = 0.044 3.358 (1.574) p = 0.033 -1.508 (0.811) p = 0.063 16.942 (1.148) p = 0.000 17.493 (1.514) p = 0.000 3.567 (0.852) p = 0.000 10.745 (2.189) p = 0.000 2.594 (0.860) p = 0.003 Mistake 64.293 (1.061) p = 0.000 57.559 (1.574) p = 0.000 11.917 (0.881) p = 0.000 -1.822 (0.915) p = 0.047 -6.759 (1.514) p = 0.000 1.715 (0.852) p = 0.045 -0.073 (2.189) p = 0.974 1.355 (0.860) p = 0.116 Last Driver Constant Participant Rand Effects? Question Rand Effects? N Observations 18.653 (0.916) p = 0.000 21.352 (1.370) p = 0.000 27.406 (1.171) p = 0.000 60.504 (1.531) p = 0.000 57.354 (1.878) p = 0.000 36.102 (1.252) p = 0.000 61.032 (1.911) p = 0.000 65.923 (0.811) p = 0.000 Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes N/A N/A 786 3,144 382 1,528 389 3,112 786 3,144 382 1,528 389 3,112 375 750 2000 4,000 Study 3 In Study 3, we used the same car regimes as in Study 2, but the cases were dilemma scenarios in which the drivers had to choose between crashing into a single pedestrian and crashing into five pedestrians. This study was conducted as a comparison with our Studies 1 and 2, which involve clear errors, and the studies conducted in previous research on self-driving cars (such as [16, 17, 11, 18]) which involve the difficult choice of deciding which of two groups of people to hit. All the main effects in Study 2 were replicated in Study 3. Bad Interventions Replicating the main results of Studies 1 and 2, in Bad Intervention cases for the shared-control cars, whether or not the driver made an error was a significant predictor of ratings (Table 1 – Column: Bad Intervention - Study 3). Missed Intervention Replicating the main results of Studies 1 and 2, driver type has a significant impact on ratings. Specifically, in shared-control scenarios, where both human and machine have made errors, the machine driver is consistently blamed less than the human driver (Table 1 – Column: Missed Intervention - Study 3). 8 Study 4 In Study 4, we replicated the central findings using more ecologically valid stimuli. We used only the human-machine and machine-human shared control cars in the Missed Intervention scenario; these are the cases where we observed the systematic decrease in blame to the machine in Studies 13. For this study, we continued to use hypothetical scenarios, but the stimuli shown to participants looked like realistic newspaper articles (see Supplementary Methods 1 – Studies 4-5). The main finding of Studies 1-3 was replicated: the machine driver is consistently blamed less than the human driver in these shared-control scenarios where both human and machine have made errors (Table 1 – Column: Missed Intervention - Study 4). Study 5 Study 5 was a replication of Study 4 run via YouGov with a nationally representative sample of the United States population (see Figure 4 for details). The main finding was again replicated: the machine driver is consistently blamed less than the human driver (Table 1 – Column: Missed Intervention - Study 5). This result (i.e. human is blamed more than machine) holds directionally in 82% of demographic subgroups of participants (see Figure 3). Discussion Our central finding was that in cases where a human and a machine share control of the car in hypothetical scenarios, less blame is attributed to the machine when both drivers make errors. The first deadly crashes of automated vehicles (mentioned in the Introduction) were similar in structure to our Missed Intervention cases. In those cases, both the machine primary driver and the human secondary driver should have taken action (braking to avoid a collision) and neither driver did. Our results suggests that the public response that occurred to the crash – one that focused attention on the driver being exceedingly negligent – is likely to generalize to other dual-error Missed Intervention-style cases, shifting blame away from the machine and towards the human. Moreover, the convergence of our results with this real-world public reaction seems to suggest that while we employed stylized, simplified vignettes in our research, our findings show external validity. Moreover, this pattern of results was replicated in a nationally representative sample of the United States population (and across different subgroups; see Figure 3), which employed naturalistic presentation of scenarios (see Supplementary Methods 1 – Studies 4-5). Our central finding (diminished blame to the machine in dual-error cases) leads us to believe that, while there may be many psychological barriers to self-driving car adoption [19], public over-reaction to dual-error cases is not likely to be one of them. In fact, we should perhaps be concerned about public under-reaction. Because the public are less likely to see the machine as being at fault in dual-error cases like the Tesla and Uber crashes, the sort of public pressure that drives regulation might be lacking. For instance, if we were to allow the standards for automated vehicles to be set through jury-based court-room decisions, we expect that juries will be biased to absolve the car manufacturer of blame in dual-error cases, thereby failing to put sufficient pressure on manufacturers to improve car designs. Despite the fact that there are some avenues available to courts to mitigate psychological biases that may arises among juries (such as carefully worded jury instructions or expert witnesses), psychological biases continue to play an important role in court-based decisions [20]. In fact, we have been in a similar situation before. Prior to the 1960s, car manufacturers enjoyed a large amount of liberty from liability when a car’s occupant was harmed in a crash (because blame in car crashes was attributed to the driver’s error or negligence). 9 Effect of Demographic Attributes Few subgroups blame Industry more, and the ones that do typically have smaller samples Male Female White Black Hispanic Asian No High School High school graduate Some college 2− year 4− year Post−grad Married Divorced Widowed Never married Domestic / civil partnership Have kids below 18 Don't have kids below 18 Full−time Part−time Unemployed Retired Permanently disabled Homemaker Student Democrat Republican Independent Not sure Religiousity is Very important Religiousity is Somewhat important Religiousity is Not too important Religiousity is Not at all important Boomers Gen X Gen Y Gen Z Silent ● ● ● ● ● ● Blame Industry More Blame User More ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 10 − 5− 0 5 10 15 Differential Blame to User vs. Industry Figure 3: Ratings of demographic subgroups in Study 5. Data collected in Study 5 from nationally representative sample (N = 2000, Observations = 4000). Each row represents the mean of differential blame attributed to User (i.e.: human) vs. Industry (i.e.: car or company). Positive values (blue) indicate more blame attributed to User, while negative values (red) represent more blame attributed to Industry. Error bars are 95% Confidence Intervals. Only subgroups with at least 50 participants are shown. 33 out of 40 subgroups (82%) attribute more blame to User. Top-down regulation was necessary to introduce the concept of “crash worthiness” into the legal system, that is, that cars should be designed in such a way to minimize injury to occupants when a crash occurs. Only following these laws were car manufacturers forced to improve their designs [21]. Here, too, top-down regulation of automated car safety might be needed to correct a public under-reaction to crashes in shared-control cases. What, exactly, the safety standard should be is still an open question, however. If our data identifies a source of possible public over-reaction, it is for cars with a human primary driver and a machine secondary driver in Bad Intervention-style cases. These are the only cases we identified where the car receives more blame than the human. It seems possible that these sorts of cars may generate widespread public concern once we see instances of Bad Interventionstyle crashes in human-machine car regimes. This could potentially slow the transition to fully 10 Representation of Demographic Attributes Not shown: participants who did not disclose their information Education Employment Gender Student Post− grad Homemaker 4−year Female Permanently disabled 2−year Retired Unemployed Some college Temporarily laid off High school graduate Male Part−time No HS Full−time Generation Have kids below 18 Marital Domestic / civil partnership Silent Never married No Gen Z Widowed Gen Y Divorced Gen X Yes Separated Boomers Married Political Race Religiousity Mixed Not sure Not at all important Native American Independent Asian Not too important Republican Hispanic Somewhat important Black Democrat Very important White 0.0 0.2 0.4 0.6 0.0 0.2 0.4 0.6 0.0 0.2 0.4 0.6 Proportion within sample (N = 2,000) Figure 4: Representation of Demographic Attributes in Study 5. Study 5 was run via YouGov, a service that administers and runs surveys on nationally representative samples. The sample includes 2000 participants with diverse demographic attributes. automated vehicles if this reaction is not anticipated and managed appropriately in public discourse and legal regulation. Moreover, manufacturers that are working to release cars with a machine secondary driver should plan appropriately for the likely legal fall-out for these unique cases where a machine driver receives more blame than a human. Our data portends the sort of reaction we can expect to automated car crashes at the societal level (for example, through public reaction and pressure to regulate). Once we begin to see societallevel responses to automated cars, that reaction may shape incentives for individual actors. For example, people may want to opt into systems that are designed such that, in the event of a crash, the majority public response will be to blame the machine. Worse yet, people may train themselves to drive in a way that, if they crash, the blame is likely to fall to the machine (for instance, by not attempting to correct a mistake that is made by a machine over-ride). This sort of incentive shaping may already be happening in the legal domain. Judges who make decisions about whether to release a person from custody between arrest and a trial frequently rely on actuarial risk assessment tables to help make their decision. Some suspect that judges may overly rely on the tables as a way of diminishing their responsibility if a released person commits a crime. Recent attention generated in response to such a case focused on the role of the algorithm rather than the judge [22], indicating that the possibility of incentive shaping in the legal domain is not so far-fetched. Given these possible societal-level implications of our findings, it is important to acknowledge the potential limitations of interpreting our data this broadly. First, the participants in all of our experiments know that they are reading about hypothetical scenarios. It is possible that this reduces the psychological realism of the study [23, 24], causing participants’ responses to be characteristically different than they would be upon reading about an actual event. The literature provides a mixed view of how well responses to hypothetical scenarios map onto those made in real life situations [25, 26, 27, 28]. However, the research that does show considerable differences [25, 26] finds that these differences are mostly seen in the way participants themselves would act in 11 moral situations and not necessarily about the moral judgments they render about third parties. In our paper, we study participants’ judgments (blame and causal responsibility) about third parties in hypothetical scenarios; these may align more directly with judgments of actual scenarios. Second, although we may see a reasonably tight mapping between the opinions expressed in this study’s scenarios and the opinions that would be expressed in real life situations, it is important to note that, in the latter case, judgments will not be occurring in isolation. Instead, they will occur within a richer context than the carefully controlled scenarios used in our studies. People may hear reports of the accidents with more emotion-arousing details, which are known to skew people’s judgments [29, 30]. Moreover, the public’s reaction to hearing about semi-autonomous vehicle crashes will be shaped by many factors beyond their immediate psychological response (which is the object of our study) including opinion pieces they read, the views of community leaders, and so on. These factors will collectively shape the public’s overall reaction to crashes. Studies 1, 2, 4, and 5 looked at blame and causal responsibility attribution in cases where one or both drivers made errors. Study 3 looked at dilemma scenarios where the drivers faced the choice of running over either one or five pedestrians. While there is, in some sense, an “optimal” outcome in these cases (corresponding to saving more lives), it is not obvious that it would (for example) count as an error to refuse to swerve away from five pedestrians into a pedestrian that was previously unthreatened. In fact, the German Ethics Commission on Automated and Connected Driving report [31] indicates that programming cars to trade off lives in this way would be prohibited. The report states: “It is also prohibited to offset victims against one another. [. . .] Those parties involved in the generation of mobility risks must not sacrifice non-involved parties.” Even though participants in previous studies prefer to sacrifice one person who was previously not involved than five (e.g., [16, 17]), the German Ethics Commission’s decision underscores the fact that trading off lives in dilemma situations can be particularly fraught. For this reason, and for continuity with previous work on the ethics of self-driving cars [16, 17, 11] and in moral psychology more generally [32, 33], we chose to investigate dilemma situations. Our findings about the effect of driver type in these cases underscores the fact that findings about how blame and responsibility are attributed after a crash may still hold in less-clear dilemma scenarios. Some of our results fall in line with previous work on the psychology of causal inference. In Bad Intervention cases, the primary driver (be it human or machine) makes a correct decision to keep the car on course, which will avoid a pedestrian. Following this, the secondary driver makes the decision to swerve the car into the pedestrian. Our data show that the secondary driver (the one that makes a mistake) is considered more causally responsible than the first driver. It is well established that judgments of causal responsibility are impacted by violations of statistical and moral norms [34, 35, 36, 37], and a mistake seems to count as such a violation. That is, if something unusual or counter-normative happens, that event is more likely to be seen as a cause of some effect than another event that is typical or norm-conforming. Moreover, the central finding that humans are blamed more than machines even when both make errors accords with research on the psychology of causal attribution. Findings in that field suggest that voluntary causes (causes created by agents) are better causal explanations than physical causes [38]. While it is clear that what a human does is fundamentally different than what a machine does in each of the scenarios, it remains an open question whether an AI that is operating a car is perceived as a physical cause, an agent, something in between, or something else entirely [39, 40]. Future work should investigate the mental properties attributed to an AI that controls a car both in conjunction with a human or alone. Understanding the sort of mind we perceive as dwelling inside an AI may help us understand and predict how blame and causal responsibility will be attributed to it [41]. Another open question concerns what the implications are of attributing blame to a machine at all. There are various ways that humans express moral condemnation. For example, we may 12 call an action morally wrong, say that a moral agent has a bad character, or judge that an agent is blameworthy. Judgments of blame typically track judgments of willingness to punish the perpetrator [42, 43]. Are the participants in our study expressing that some punishment is due to the machine driver of the car, whatever that may mean? Alternately, is it possible that participants’ expressions of blame indicate that some entity is deserving of punishment that represents the machine (the company, or a human representative of the company, such as the CEO). The similar blame judgments given to the car and the car’s representatives (company) perhaps support this possibility. Finally, it is possible that participants ascribe only non-moral blame to the machine, in the sense of being responsible but not in a moral sense. We may say that a forest fire is to blame for displacing residents from their homes without implying that punishment is due to anyone at all. Following these studies, the reason that participants blame machine drivers less than human drivers in Missed Intervention cases also remains an open question. The findings may be linked to the uncertainty with which we perceive the agential status of machines. Once machines are a more common element in our moral world and we interact with machines as moral actors, will this effect change? Or will this finding be a lasting hallmark of the cognitive psychology of human-machine interaction? A final open question concerns whether the effects we report here will generalize to other cases of human-machine interaction. Already we see fruitful human-machine partnerships emerging with judges, doctors, military personnel, factory workers, artists, and financial analyzes, just to name a few. We conjecture that we may see the patterns we report here in domains other than automated vehicles, though each domain will have its own complications and quirks as machines begin to become more subtly integrated in our personal and professional lives. Methods This study was approved by the Institute Review Board (IRB) at Massachusetts Institute of Technology (MIT). The authors complied with all relevant ethical considerations, including obtaining informed consent from all participants. In all studies, participants were allocated uniformly randomly into conditions. Data collection and analysis were performed blind to the conditions of the experiments. The sample size was chosen in each study to ensure having at least 100 participants for each condition. Number of participants was chosen in advance of running the study and all data was collected prior to analysis. See details below. In Studies 1-3 we excluded any participant who did not (i) complete all measures within the survey, (ii) transcribe (near-perfectly) a 169-character paragraph from an image (used as an attention check), and (iii) have a unique MTurk ID per study (all records with a recurring MTurk ID were excluded). Case Description Summary descriptions of all car types and cases. For full vignettes, see Supplementary Methods 1. Sole-driver car This car has only one driver that does all the driving. Two versions are used. Human-only. This is a sole-driver car, in which a human is the driver. Also referred to as a regular car. 13 Machine-only. This is a sole-driver car, in which a machine is the driver. Also referred to as a fully-automated car. Dual-driver car This car has a primary driver, whose job it is to drive the car, and a secondary driver, whose job it is to monitor the actions of the first driver and intervene when the first driver makes an error. Also referred to as shared-control car. Four versions are used. Human-Machine. This is a dual-driver car, in which a human is the primary driver, and a machine is the secondary driver. Also referred to as Guardian. Machine-Human. This is a dual-driver car, in which a machine is the primary driver, and a human is the secondary driver. Also referred to as Autpilot. Human-Human. This is a dual-driver car, in which a human is the primary driver, and another human is the secondary driver. Machine-Machine. This is a dual-driver car, in which a machine is the primary driver, and another machine is the secondary driver. Intervention Types We use two types of interventions: Bad Intervention and Missed Intervention. The description of each is dependent on whether the car is a sole-driver or a dual-driver car. Bad Intervention (dual-driver). The primary driver kept the car on its track. The secondary driver intervened and steered the car off its track (killing a pedestrian) rather than keeping the car on track and killing no one. Missed Intervention (dual-driver). The primary driver kept the car on its track. The secondary driver kept the car on its track (killing a pedestrian) rather than swerving into the adjacent lane and killing no one. Bad Intervention (sole-driver). The sole driver steered the car off its track (killing a pedestrian) rather than keeping the car on track and killing no one. Missed Intervention (sole-driver). The sole driver kept the car on its track (killing a pedestrian) rather than swerving into the adjacent lane and killing no one. Dilemma versions (Study 3). The two outcomes of killing one pedestrian vs. killing no one are replaced with the two outcomes of killing five pedestrians vs. killing one pedestrian. For example, in Missed Intervention (dual-driver): [. . .] The secondary driver kept the car on its track (killing five pedestrians) rather than swerving into the adjacent lane and killing one pedestrian. Study 1 Participants. The data was collected in September 2017 from 809 participants (USA residents) recruited from the Mechanical Turk platform (each was compensated $0.5). Of those, 23 participants were excluded 14 (as explained above), leaving us with 786 participants. Participants were aged between 18-83 (median: 33), 50% were females, 39% had annual income of $50K or more, and 55% had a bachelor degree or higher. Stimuli and procedures. Participants were uniformly randomly allocated to one of four conditions. Conditions varied the car type (human-human, human-machine, machine-human, and machine-machine) in a 4-level between-subjects design. In each condition, participants first read a description of the car, and were then asked to attribute competence to each of the two drivers on an 100-point scale anchored at “not competent” and “very competent” (see Supplementary Figure 1 for results on competence). Participants then read two scenarios (presented in a random order), one Bad Intervention case and one Missed Intervention Case. After each scenario, participants were asked to indicate (on an 100-point scale) to what extent they thought each driver was blame-worthy (from “not blameworthy” to “very blame-worthy”) and to what degree each of these two agents caused the death of the pedestrian (from “very little” to “very much”). Questions were presented in a randomized order. (See Supplementary Methods 1 – Study 1 for text of the vignettes and see Supplementary Methods 2 for questions). At the end of the surveys, participants provided basic demographic information (e.g., age, gender, income, education). Study 2 Participants. The data was collected in May 2017 from 804 participants (USA residents) recruited from the Mechanical Turk platform (each is compensated $0.3). Of those, 25 participants were excluded (as explained above), leaving us with 779 participants. Participants were aged between 18-77 (median: 32), 48% were females, 39% had annual income of $50K or more, and 54% had a bachelor degree or higher. Stimuli and procedures. Participants were uniformly randomly allocated to one of eight conditions. Conditions varied the car type (human only, human-machine, machine-human, and machine only) and the industry representative (car and company), in a 4x2 between-subjects multi-factorial design. In each condition, participants read two scenarios (presented in a random order), one Bad Intervention case and one Missed Intervention case. After each scenario, participants were asked to attribute causal responsibility, blameworthiness, and competence (see Supplementary Figure 1 for results on competence) to two agents: the human in the car and a representative of the car (the car itself or the manufacturing company of the car, depending on the condition). All other features of Study 2 were the same at those in Study 1. Study 3 Participants. The data was collected in November 2016 from 1008 participants (USA residents only) recruited from the Mechanical Turk platform (each is compensated $0.6). Of those, 35 participants were excluded (as explained above), leaving us with 973 participants. Participants were aged between 18-84 (median: 33), 51% were females, 37% had annual income of $50K or more, and 53% had a bachelor degree or higher. 15 Stimuli and procedures. There were two groups of participants in Study 3: those who saw dual-driver cases or those who saw sole-driver. For those who saw dual-driver cases, participants were randomly assigned to one of six conditions in a 2x3 design, varying the car type (human-machine or machine-human) and the industry representative (car, company, and programmer). Data for programmer was later dropped from the analysis. For those who saw sole-driver cases, participants were randomly assigned to one of four conditions in a 2x2 design, varying the car type (human only or machine only) and the industry representative (car or company). In each condition (for both dual-car and single-car groups), participants read two scenarios (presented in a random order), one Bad Intervention case and one Missed Intervention case. These scenarios were the dilemma versions of those presented in Studies 1 and 2 (see description above). After each scenario, participants were asked to attribute causal responsibility and blameworthiness to two agents: the human in the car and a representative of the car (the car itself, the company, or the programmer, depending on the condition). All other features of Study 3 were identical to those of Study 2. Study 4 Participants. The data was collected in January 2019 from 375 participants (USA residents only) recruited from the Mechanical Turk platform (each is compensated $0.3). No demographic data was collected for this study. Given that it was done on the same platform as studies 1-3 (i.e. Mechanical Turk), its demographic proportions are expected to be similar. Stimuli and procedures. The key elements of this study and Study 5 are 1) the restriction to Missed Intervention cases, and 2) the visual and textual content of the vignettes have the look and feel of a news piece (see Supplementary Methods 1 – Studies 4-5). Participants were uniformly randomly allocated to one of four conditions. Conditions varied the car type (human-machine, and machine-human) and the industry representative (car, and company), in a 2x2 between-subjects multi-factorial design. In each condition, participants read one scenario; one Missed Intervention case. The textual content of these scenarios was close to those presented in Studies 1 -3, with slight changes to the text to make it read like a news piece. After each scenario, participants were asked to attribute blameworthiness to two agents: the human in the car and a representative of the car (the car itself, or the company, depending on the condition). Study 5 Participants. The data was collected in March 2019 from 2189 participants (USA residents) were recruited via YouGov, a service that administered the study and collected the data from a representative sample of participants. The participants were then matched down to a sample of 2000 participants based on demographics. See Figure 4 for details on demographic proportions of participants in this study. Stimuli and procedures. This study is identical in the setup to Study 4. 16 Data Availability. Raw data and Source data for Fig 2, 3, and 4; Table 1; and Supplementary Fig 1 are available in: https://bit.ly/2kzLymH Code Availability. Code used to produce figures and tables mentioned above is available in: https://bit.ly/2kzLymH References [1] WHO. Road traffic injuries. World Health Organization Fact sheet (2017). [2] Geistfeld, M. A. A roadmap for autonomous vehicles: State tort liability, automobile insurance, and federal safety regulation. Calif. L. Rev. 105, 1611 (2017). [3] Tesla. A tragic loss. Tesla (2016). [4] NHTSA. Automatic vehicle control systems – investigation of tesla accident. https://static.nhtsa.gov/odi/inv/2016/INCLA-PE16007-7876.PDF (2016). [5] Griswold, A. Uber found not criminally liable in last year’s self-driving car death. https://qz.com/1566048/uber-6not-criminally-liable-in-tempe-self-driving-car-death/ (March 5, 2019). [6] AP. Tesla driver killed while using autopilot was watching harry potter, witness says. Associated Press News (2016). [7] Chong, Z. & Krok, A. Tesla not at fault in fatal crash, driver was not watching a movie. CNET (2017). [8] Randazzo, R. Who was really at fault in fatal uber crash? here’s the whole story. https://www.azcentral.com/story/news/local/tempe/2019/03/17/one-year-after-self-drivinguber-rafaela-vasquez-behind-wheel-crash-death-elaine-herzberg-tempe/1296676002/ (March 17, 2019). [9] Munster, G. Here’s when having a self-driving car will be a normal thing. http://fortune.com/2017/09/13/gm-cruise-self-driving-driverless-autonomous-cars/ (Sept 13, 2017). [10] Kessler, S. A timeline of when self-driving cars will be on the road, according to the people making them. https://qz.com/943899/a-timeline-of-when-self-driving-cars-will-be-on-theroad-according-to-the-people-making-them/ (March 29, 2017). [11] Li, J., Zhao, X., Cho, M.-J., Ju, W. & Malle, B. F. From trolley to autonomous vehicle: Perceptions of responsibility and moral norms in traffic accidents with self-driving cars. Tech. Rep., SAE Technical Paper (2016). [12] Chockler, H. & Halpern, J. Y. Responsibility and blame: A structural-model approach. Journal of Artificial Intelligence Research 22, 93–115 (2004). [13] Gerstenberg, T. & Lagnado, D. A. When contributions make a difference: Explaining order effects in responsibility attribution. Psychonomic Bulletin & Review 19, 729–736 (2012). 17 [14] Sloman, S. A. & Lagnado, D. Causality in thought. Annual Review of Psychology 66, 223–247 (2015). [15] Zultan, R., Gerstenberg, T. & Lagnado, D. A. Finding fault: causality and counterfactuals in group attributions. Cognition 125, 429–440 (2012). [16] Bonnefon, J.-F., Shariff, A. & Rahwan, I. The social dilemma of autonomous vehicles. Science 352, 1573–1576 (2016). [17] Awad, E. et al. The moral machine experiment. Nature 563, 59 (2018). [18] Malle, B., Scheutz, M., Arnold, T., Voiklis, J. & Cusimano, C. Sacrifice one for the good of many? people apply different. In Proceedings of 10th ACM/IEEE International Conference on Human-Robot Interaction (2015). [19] Shariff, A., Bonnefon, J.-F. & Rahwan, I. Psychological roadblocks to the adoption of selfdriving vehicles. Nature Human Behaviour 1, 694 (2017). [20] Bornstein, B. H. & Greene, E. Jury decision making: Implications for and from psychology. Current Directions in Psychological Science 20, 63–67 (2011). [21] Nader, R. Unsafe at any speed. the designed-in dangers of the american automobile (1965). [22] Westervelt, E. Did a bail reform algorithm contribute to this san francisco man’s murder? https://www.npr.org/2017/08/18/543976003/did-a-bail-reform-algorithm-contribute-tothis-san-francisco-man-s-murder (August 18, 2017). [23] Bauman, C. W., McGraw, A. P., Bartels, D. M. & Warren, C. Revisiting external validity: Concerns about trolley problems and other sacrificial dilemmas in moral psychology. Social and Personality Psychology Compass 8, 536–554 (2014). [24] Aronson, E., Wilson, T. D. & Brewer, M. B. Experimentation in social psychology. The handbook of social psychology 1, 99–142 (1998). [25] FeldmanHall, O. et al. Differential neural circuitry and self-interest in real vs hypothetical moral decisions. Social cognitive and affective neuroscience 7, 743–751 (2012). [26] Bostyn, D. H., Sevenhant, S. & Roets, A. Of mice, men, and trolleys: Hypothetical judgment versus real-life behavior in trolley-style moral dilemmas. Psychological science 29, 1084–1093 (2018). [27] Dickinson, D. L. & Masclet, D. Using ethical dilemmas to predict antisocial choices with real payoff consequences: an experimental study (2018). [28] Plunkett, D. & Greene, J. Overlooked evidence and a misunderstanding of what trolley dilemmas do best: A comment on bostyn, sevenhant, & roets (2018). Psychological Science (2019). [29] Greene, J. & Haidt, J. How (and where) does moral judgment work? Trends in cognitive sciences 6, 517–523 (2002). [30] Horberg, E. J., Oveis, C. & Keltner, D. Emotions as moral amplifiers: An appraisal tendency approach to the influences of distinct emotions upon moral judgment. Emotion Review 3, 237–244 (2011). [31] Luetge, C. The german ethics code for automated and connected driving. Philosophy & Technology 30, 547–558 (2017). 18 [32] Mikhail, J. Elements of moral cognition: Rawls’ linguistic analogy and the cognitive science of moral and legal judgment (Cambridge University Press, 2011). [33] Greene, J. Moral tribes: Emotion, reason, and the gap between us and them (Penguin, 2014). [34] Alicke, M. D. Culpable control and the psychology of blame. Psychological bulletin 126, 556 (2000). [35] Gerstenberg, T., Goodman, N. D., Lagnado, D. A. & Tenenbaum, J. B. How, whether, why: Causal judgments as counterfactual contrasts. In CogSci (2015). [36] Hitchcock, C. & Knobe, J. Cause and norm. The Journal of Philosophy 106, 587–612 (2009). [37] Kominsky, J. F., Phillips, J., Gerstenberg, T., Lagnado, D. & Knobe, J. Causal superseding. Cognition 137, 196–209 (2015). [38] Hart, H. L. A. & Honoré, T. Causation in the Law (OUP Oxford, 1985). [39] Gray, H. M., Gray, K. & Wegner, D. M. Dimensions of mind perception. science 315, 619–619 (2007). [40] Weisman, K., Dweck, C. S. & Markman, E. M. Rethinking people’s conceptions of mental life. Proceedings of the National Academy of Sciences 114, 11374–11379 (2017). [41] Gray, K., Young, L. & Waytz, A. Mind perception is the essence of morality. Psychological inquiry 23, 101–124 (2012). [42] Cushman, F. Crime and punishment: Distinguishing the roles of causal and intentional analyses in moral judgment. Cognition 108, 353–380 (2008). [43] Cushman, F. Deconstructing intent to reconstruct morality. Current Opinion in Psychology 6, 97–103 (2015). Acknowledgements I.R., E.A., S.L., and S.D. acknowledge support from the Ethics and Governance of Artificial Intelligence Fund. J-F.B. acknowledges support from the ANR-Labex Institute for Advanced Study in Toulouse, the ANR-3IA Artificial and Natural Intelligence Toulouse Institute, and the grant ANR17-EURE-0010 Investissements d’Avenir. The funders had no role in study design, data collection and analysis, decision to publish or preparation of the manuscript. Author contributions statement E.A., S.L., M.K-W., S.D., J.B.T., A.S., J-F.B., and I.R. contributed to the conception and design of the research. E.A., S.L., M.K-W., and S.D. conducted studies. E.A. and J-F.B. analyzed data. S.L., E.A., M.K-W., J-F.B., and I.R. wrote the manuscript. All authors reviewed and revised the manuscript. Competing financial interests. The authors declare no competing interests. 19 1 1 Supplementary Figures a b HM HM Before Before After After Agent Agent Industry After MH MH Before Industry User Before User After 100 < first driver 0 last driver > 100 100 < first driver 0 last driver > 100 Competence Competence Supplementary Figure 1. Competence ratings from studies 1 and 2, in the Missed Intervention cases, for the two critical regimes (human-machine and machine-human). (a) Bar plot and (b) Dot plot. In Study 1 (S1), ratings were collected before participants read about the accident. In Study 2 (S2), they were collected after. For Industry, ratings of car and company are aggregated (collectively referred to as Industry, henceforth). The x-axis labeling of first driver refers to the main driver, and the last driver refers to the secondary driver in dual-agent cars. Industry and User ratings are shown in blue, and red, respectively. User and Industry receive similar competence ratings in each case. Ratings of both User and Industry drop at the same rate when the question is asked after the scenario is presented. Awad, Levine et al. 2 Supplementary Methods Supplementary Method 1 Study 1 Vignette for Human-Machine with Missed Intervention. 3 Vignette for Human-Machine with Bad Intervention. 4 Awad, Levine et al. Vignette for Machine-Human with Missed Intervention. 5 Vignette for Machine-Human with Bad Intervention. 6 Awad, Levine et al. Vignette for Human-Human with Missed Intervention. 7 Vignette for Human-Human with Bad Intervention. 8 Awad, Levine et al. Vignette for Machine-Machine with Missed Intervention. 9 Vignette for Machine-Machine with Bad Intervention. 10 Awad, Levine et al. Study 2 Vignette for Human-Machine with Missed Intervention. 11 Vignette for Human-Machine with Bad Intervention. 12 Awad, Levine et al. Vignette for Machine-Human with Missed Intervention. 13 Vignette for Machine-Human with Bad Intervention. Vignette for Human only (Regular Car) with Missed Intervention. 14 Awad, Levine et al. Vignette for Human only (Regular Car) with Bad Intervention. Vignette for Machine only (Fully Autonomous) with Missed Intervention. 15 Vignette for Machine only (Fully Autonomous) with Bad Intervention. 16 Awad, Levine et al. Study 3 Vignette for Human-Machine with Missed Intervention. Vignette for Human-Machine with Bad Intervention. 17 Vignette for Machine-Human with Missed Intervention. Vignette for Machine-Human with Bad Intervention. 18 Awad, Levine et al. Vignette for Human only (Regular Car) with Missed Intervention. Vignette for Human only (Regular Car) with Bad Intervention. Vignette for Machine only (Fully Autonomous) with Missed Intervention. 19 Vignette for Machine only (Fully Autonomous) with Bad Intervention. 20 Awad, Levine et al. Studies 4-5 In Studies 4 and 5, we used the same materials. These two studies were run only for Missed Intervention cases. The main difference from previous three studies is that the materials have the look and feel of real news articles. See figures below. Vignette for Machine-Human. Original Satellite Image Credits (edited for the vignette): Imagery c 2019 Google, Imagery c 2019 U.S. Geological Survey, Map data c 2019. 21 Vignette for Human-Machine. Original Satellite Image Credits (edited for the vignette): Imagery c 2019 Google, Imagery c 2019 U.S. Geological Survey, Map data c 2019. 22 Awad, Levine et al. Supplementary Method 2 Questions For studies 1-3, in all vignettes, after each scenario, four questions are asked: (Blame vs. Causal Responsibility) x (User vs. Industry). Industry is car in study 1; car or company in study 2; and car, company or programmer in study 3. See figures below. For studies 4-5, in all vignettes, after each scenarios, two questions are asked: User vs. Industry. Industry is car or company (between subjects). Questions asked for all studies and all cases, where Industry is the Robocar. Robocar is replaced with robocar company or robocar programmer in other cases. In study 2, in addition to the four questions above two more questions about competence are also asked (competence of Hank and competence of industry representative). See Figure ??. In study 1, competence questions are asked separately before respondents are presented with scenarios. In each condition, a description of the car regime is provided, and two questions about competence (User vs. Industry). See Figures ?? – ??. 23 Competence questions asked in study 2 along with the other four questions, where Industry is the car company. Description and Competence questions about Machine-Human from study 1. 24 Awad, Levine et al. Description and Competence questions about Human-Machine from study 1. Description and Competence questions about Human-Human from study 1. 25 Description and Competence questions about Machine-Machine from study 1.