Abstract
In this paper, I develop a preliminary framework that permits groups (or ‘systems’) to be moral agents. I show that this has advantages over traditional accounts of moral agency when applied to cases where machines are involved in moral actions. I appeal to two thought experiments to show that the traditional account can lead us to counterintuitive consequences. Then I present what I call the ‘systematic account’ which I argue avoids these counterintuitive consequences. On my account, machines can be partial moral agents currently fulfilling some but not all of the conditions required for moral agency. Thus, when a machine is part of a group of agents, it can be part of a system that is a moral agent. This framework is a useful starting point as it preserves aspects of traditional accounts of moral agency while also including machines in our moral deliberations.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
This is in contrast to accounts (such as Floridi and Sanders (2004)) that break with the requirements of TA. I am not wedded to having to preserve parts of TA. However, since many philosophers seem committed to TA generally capturing some central things about agency, I think building on TA is a good starting point.
References
Clark, A., Chalmers, D.: The extended mind. Analysis 58(1), 7–19 (1998)
Floridi, L., Sanders, J.W.: On the morality of artificial agents. Minds Mach. 14(3), 349–379 (2004)
Himma, K.E.: Artificial agency, consciousness, and the criteria for moral agency: what properties must an artificial agent have to be a moral agent? Ethics Inf. Technol. 11(1), 19–29 (2009)
Johnson, D.G.: Computer systems: Moral entities but not moral agents. Ethics Inf. Technol. 8, 195–204 (2006)
Moor, J.H.: The nature, importance, and difficulty of machine ethics. IEEE Intell. Syst. 21(4), 18–21 (2006)
Parthemore, J., Whitby, B.: What makes any agent a moral agent? Reflections on machine consciousness and moral agency. Int. J. Mach. Conscious. 5(02), 105–129 (2013)
Searle, J.R.: Minds, brains, and programs. Behav. Brain Sci. 3(3), 417–424 (1980)
Torrance, S.: Ethics and consciousness in artificial agents. AI Soc. 22(4), 495–521 (2008)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Hardalupas, M. (2018). A Systematic Account of Machine Moral Agency. In: Müller, V. (eds) Philosophy and Theory of Artificial Intelligence 2017. PT-AI 2017. Studies in Applied Philosophy, Epistemology and Rational Ethics, vol 44. Springer, Cham. https://doi.org/10.1007/978-3-319-96448-5_27
Download citation
DOI: https://doi.org/10.1007/978-3-319-96448-5_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-96447-8
Online ISBN: 978-3-319-96448-5
eBook Packages: Computer ScienceComputer Science (R0)