Rage against the authority machines: how to design artificial moral advisors for moral enhancement

2Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper aims to clear up the epistemology of learning morality from artificial moral advisors (AMAs). We start with a brief consideration of what counts as moral enhancement and consider the risk of deskilling raised by machines that offer moral advice. We then shift focus to the epistemology of moral advice and show when and under what conditions moral advice can lead to enhancement. We argue that people’s motivational dispositions are enhanced by inspiring people to act morally, instead of merely telling them how to act. Drawing upon these insights, we claim that if AMAs are to genuinely enhance people morally, they should be designed as inspiration and not authority machines. In the final section, we evaluate existing AMA models to shed light on which holds the most promise for helping to make users better moral agents.

Cite

CITATION STYLE

APA

Landes, E., Voinea, C., & Uszkai, R. (2025). Rage against the authority machines: how to design artificial moral advisors for moral enhancement. AI and Society, 40(4), 2237–2248. https://doi.org/10.1007/s00146-024-02135-3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free