Skip to main content
Attributions of Morality and Mind to Artificial Intelligence after Real-World Moral Violations
Computers in Human Behavior
  • Daniel Burton Shank, Missouri University of Science and Technology
  • Alyssa DeSanti

The media has portrayed certain artificial intelligence (AI) software as committing moral violations such as the AI judge of a human beauty contest being “racist” when it selected predominately light-skinned winners. We examine people's attributions of morality for seven such real-world events that were first publicized in the media, experimentally manipulating the occurrence of a violation and the inclusion of information about the AI's algorithm. Both the presence of the moral violation and the information about the AI's algorithm increase participant's reporting of a moral violation occurring in the event. However, even in the violation outcome conditions only 43.5 percent of the participants reported that they were sure that a moral violation occurred. Addressing whether the AI is blamed for the moral violation we found that people attributed increased wrongness to the AI -- but not to the organization, programmer, or users -- after a moral violation. In addition to moral wrongness, the AI was attributed moderate levels of awareness, intentionality, justification, and responsibility for the violation outcome. Finally, the inclusion of the algorithm information marginally increased perceptions of the AI having mind, and perceived mind was positively related to attributions of intentionality and wrongness to the AI.

Psychological Science
Keywords and Phrases
  • Algorithm,
  • Artificial intelligence,
  • Attributions,
  • Morality,
  • Perceived mind,
  • Responsibility
Document Type
Article - Journal
Document Version
File Type
© 2018 Elsevier, All rights reserved.
Publication Date
Citation Information
Daniel Burton Shank and Alyssa DeSanti. "Attributions of Morality and Mind to Artificial Intelligence after Real-World Moral Violations" Computers in Human Behavior Vol. 86 (2018) p. 401 - 411 ISSN: 0747-5632
Available at: