Abstract
A lot has been written recently on the ineffectiveness of the current plethora of AI ethics regulations available in the public and private sectors. I approach this concern from a novel angle by critically reflecting from within the ethics of technology on current AI ethics discourse, which is mostly still deeply Cartesian, especially when it comes to policy-making. I start with an analysis of current AI ethics vocabulary and point to its value-laden and Cartesian nature. In a first step towards moving away from Cartesianism I then briefly take the reader on a journey through pertinent aspects of trans-human discourse as illustrated by Clark’s proposal of human minds as ‘extended’. I then consider Verbeek and Kudina’s work in post-phenomenological mediation theory to enrich Clark’s suggestions by acknowledging a more active role for technology in co-shaping humans and their socio-cultural worlds. As a result, via a novel notion of ‘extended moral agency’, I define a notion of ‘moral affordance’ to inform a new non-Cartesian tradition for AI ethics discourse and policy-making. Finally, I briefly comment on implications of my argument for the future of AI ethics regulation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Asaro, P.M.: What should we want from a robot ethic? Int. Rev. Inform. Ethics 6, 9–16 (2006)
Bhatt, U., et al.: Explainable machine learning in deployment. In: Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, pp. 648–657 (2020)
Boddington, P.: AI and moral thinking: how can we live well with machines to enhance? AI Ethics 1, 109–111 (2021). https://doi.org/10.1007/s43681-020-00017-0
Borgmann, A.: Technology and the Character of Contemporary Life: A Philosophical Inquiry. University of Chicago Press, Chicago (1984)
Borgmann, A.: Holding on to Reality. The Nature of Information at the Turn of the Millennium. University of Chicago Press, Chicago (1999)
Bostrom, N.: Superintelligence Paths, Dangers, Strategies. Oxford University Press, Oxford (2014)
Bostrom, N., Yudkowsky, E.: The ethics of artificial intelligence. In: The Cambridge Handbook of Artificial Intelligence, pp. 316–334 (2014)
Brundage, M.: Limitations and risks of machine ethics. J. Exp. Theor. Artif. Intell. 26(3), 355–372 (2014)
Capurro, R.: Digital hermeneutics: an outline. AI Soc. 25(1), 35–42 (2010). https://doi.org/10.1007/s00146-009-0255-9
Cassam, Q.: Vices of the Mind: From the Intellectual to the Political. Oxford University Press, Oxford (2019)
Clark, A.: Natural-Born Cyborgs: Minds, Technologies, and the Future of Human Intelligence. Oxford University Press, Oxford (2004)
Crawford, K.: The Trouble with Bias. NIPS 2017 Keynote. https://www.youtube.com/watch?v=fMym_BKWQzk
Crawford, K., Calo, R.: There is a blind spot in AI research. Nat. News 538(7625), 311–313 (2016)
de Boer, B., Te Molder, H., Verbeek, P.-P.: The perspective of the instruments: mediating collectivity. Found. Sci. 23(4), 739–755 (2018). https://doi.org/10.1007/s10699-018-9545-3
Dennett, D.C.: Consciousness Explained. Penguin Books, London (1991)
Dreyfus, H.L.: What Computers Can’t Do: A Critique of Artificial Reason. Harper & Row, New York (1972)
Dreyfus, H.L.: On the Internet. Routledge, New York (2001)
Druga, S., Williams, R., Breazeal, C., Resnick, M.: “Hey Google is it OK if i eat you?” Initial explorations in child-agent interaction. In: Proceedings of the 2017 Conference on Interaction Design and Children, pp. 595–600. Association for Computing Machinery, New York (2017)
European Union Commission: Ethics Guidelines for Trustworthy AI. https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai
Feenberg, A.: What is philosophy of technology? In: Defining Technological Literacy, pp. 5–16. Palgrave Macmillan (2006)
Floridi, L., Sanders, J.W.: On the morality of artificial agents. Minds Mach. 14(3), 349–379 (2004). https://doi.org/10.1023/B:MIND.0000035461.63578.9d
Gadamer, H.G.: Philosophical Hermeneutics. University of California Press, Berkeley (1977). (Linge, D.E Trans.)
Gibson, J.J.: The Ecological Approach to Visual Perception, Classic Psychology Press Ltd., New York (2014)
Hagendorff, T.: The ethics of AI ethics: an evaluation of guidelines. Minds Mach. 30(1), 99–120 (2020). https://doi.org/10.1007/s11023-020-09517-8
Haselager, P., Mecacci, G.: Superethics instead of superintelligence: know thyself, and apply science accordingly. AJOB Neurosci. 11(2), 113–119 (2014)
Ihde, D.: Technics and Praxis: A Philosophy of Technology. D. Reidel Publishing Company, Boston (1979)
Ihde, D.: Technology and the Lifeworld: From Garden to Earth. Indiana University Press, Bloomington (1990)
Jesus, S., et al.: How can i choose an explainer? An application-grounded evaluation of post-hoc explanations. In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 805–815 (2021)
Jobin, A., Ienca, M., Vayena, E.: The global landscape of AI ethics guidelines. Nat. Mach. Intell. 1(9), 389–399 (2019). https://doi.org/10.1038/s42256019-0088-2
Kruger, J.: Nature, culture, AI and the common good – considering AI’s place in Bruno Latour’s politics of nature. In: Gerber, A. (ed.) SACAIR 2021. CCIS, vol. 1342, pp. 21–33. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-66151-9_2
Kudina, O.: “Alexa, who am I?’’: voice assistants and hermeneutic lemniscate as the technologically mediated sense-making. Hum. Stud. 44(2), 233–253 (2021). https://doi.org/10.1007/s10746-021-09572-9
Kudina, O., Verbeek, P.P.: Ethics from within: Google Glass, the Collingridge dilemma, and the mediated value of privacy. Sci. Technol. Hum. Values 44(2), 291–314 (2019)
Latour, B.: We Have Never Been Modern. Harvard University Press, Cambridge (1993)
Latour, B.: On actor-network theory: a few clarifications. Soziale Welt 47(4), 369–381 (1996)
Lidskog, R., Soneryd, L., Uggla, Y.: Transboundary Risk Governance. Earthscan, London (2009)
McCarthy, J., Wright, P.: Technology as Experience. MIT Press, Cambridge (2004)
Mellamphy, N.B.: Humans in the loop. Human-centrism, Posthumanism, and AI. Nat. Cult. 16(1), 11–27 (2021). https://www.berghahnjournals.com/view/journals/nature-and-culture/16/1/nc160102.xml
Mittelstadt, B.: Principles alone cannot guarantee ethical AI. Nat. Mach. Intell. 1(11), 501–507 (2019). https://doi.org/10.1038/s42256-019-0114-4
Moor, J.: The nature, importance, and difficulty of machine ethics. IEEE Intell. Syst. 21(4), 18–21 (2006)
Rosenberger, R., Verbeek, P.P.: A field guide to postphenomenology. In: Rosenberger, R., Verbeek, P.P. (eds.) Postphenomenological Investigations: Essays on Human-Technology Relations, pp. 7–42. Lexington Books (2015)
Russell, S.: Human Compatible - Artificial Intelligence and the Problem of Control. Viking, London (2019)
Selbst, A.D., Boyd, D., Friedler, S.A., Venkatasubramanian, S., Vertesi, J.: Fairness and abstraction in sociotechnical systems. In: Proceedings of the Conference on Fairness, Accountability, and Transparency, FAT* 2019, pp. 59–68. Association for Computing Machinery, New York (2019). https://doi.org/10.1145/3287560.3287598
Selinger, E.: Confronting the moral dimensions of technology through mediation theory. Philos. Technol. 27, 287–313 (2014). https://doi.org/10.1007/s13347-011-0054-3
Swyngedouw, E.: Impossible ‘sustainability’ and the postpolitical condition. In: Krüger, R., Gibbs, D. (eds.) The Sustainable Development Paradox, pp. 13–40. Guilford Press (2007)
Turkle, S.: The Second Self: Computers and the Human Spirit. Simon & Schuster Inc., New York (1984)
Uggla, Y.: What is this thing called ‘natural’? The nature-culture divide in climate change and biodiversity policy. J. Polit. Ecol. 17(1), 79–99 (2010). https://doi.org/10.2458/v17i1/21701
UNESCO: Preliminary Report on the First Draft of the Recommendation on the Ethics of Artificial Intelligence. https://unesdoc.unesco.org/ark:/48223/pf0000374266
Vallor, S.: Technology and the Virtues: A Philosophical Guide to a Future Worth Wanting. Oxford University Press, Oxford (2016)
Veale, M., Binns, R.: Fairer machine learning in the real world: mitigating discrimination without collecting sensitive data. Big Data Soc. 4(2) (2017). https://doi.org/10.1177/2053951717743530
Verbeek, P.P.: What Things Do: Philosophical Reflections on Technology, Agency, and Design. Pennsylvania State University Press, University Park (2005)
Verbeek, P.P.: Obstetric ultrasound and the technological mediation of morality: a post-phenomenological analysis. Hum. Stud. 31(1), 11–26 (2008). https://doi.org/10.1007/s10746-007-9079-0
Verbeek, P.P.: Moralizing Technology: Understanding and Designing the Morality of Things. University of Chicago Press, Chicago (2011)
Wallach, W., Allen, C.: Moral Machines: Teaching Robots Right from Wrong. Oxford University Press, Oxford (2009)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Ruttkamp-Bloem, E. (2022). Re-imagining Current AI Ethics Policy Debates: A View from the Ethics of Technology. In: Jembere, E., Gerber, A.J., Viriri, S., Pillay, A. (eds) Artificial Intelligence Research. SACAIR 2021. Communications in Computer and Information Science, vol 1551. Springer, Cham. https://doi.org/10.1007/978-3-030-95070-5_21
Download citation
DOI: https://doi.org/10.1007/978-3-030-95070-5_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-95069-9
Online ISBN: 978-3-030-95070-5
eBook Packages: Computer ScienceComputer Science (R0)