Abstract
The rapid deployment of semi-autonomous systems (i.e., systems requiring human monitoring such as Uber AVs) poses ethical challenges when these systems face morally-laden situations. We ask how people evaluate morally-laden decisions of humans who monitor these systems in situations of unavoidable harm. We conducted three pre-registered experiments (total N = 1811), using modified trolley moral dilemma scenarios. Our findings suggest that people apply different criteria when judging morality and deserved punishment of regular-car versus AV drivers. Regular-car drivers are judged according to a consequentialist minimizing harm criterion, whereas AV drivers are judged according to whether or not they took action, with a more favorable prior for acting. Integrating judgment and decision-making research with moral psychology, the current research illuminates how the presence versus absence of automation affects moral judgments.
Original language | English |
---|---|
Article number | 107590 |
Journal | Computers in Human Behavior |
Volume | 140 |
DOIs | |
State | Published - 1 Mar 2023 |
ASJC Scopus subject areas
- Arts and Humanities (miscellaneous)
- Human-Computer Interaction
- General Psychology