Abstract
The fact that real-world decisions made by artificial intelligences (AI) are often ethically loaded has led a number of authorities to advocate the development of “moral machines”. I argue that the project of building “ethics” “into” machines presupposes a flawed understanding of the nature of ethics. Drawing on the work of the Australian philosopher, Raimond Gaita, I argue that ethical dilemmas are problems for particular people and not (just) problems for everyone who faces a similar situation. Moreover, the force of an ethical claim depends in part on the life history of the person who is making it. For both these reasons, machines could at best be engineered to provide a shallow simulacrum of ethics, which would have limited utility in confronting the ethical and policy dilemmas associated with AI.
Original language | English |
---|---|
Pages (from-to) | 685-693 |
Number of pages | 9 |
Journal | AI and Society |
Volume | 36 |
Issue number | 3 |
DOIs | |
Publication status | Published - Sep 2021 |
Keywords
- Artificial intelligence
- Ethics
- Machine ethics
- Moral authority
- Raimond Gaita