- Alex Hern. 2017. Give robots ‘personhood’ status. The Guardian: EU committee argues. (Jan. 2017). Retrieved January 12, 2020 from https://www.theguardian.com/technology/2017/jan/12/give-robots-personhood-status-eu-committee-arguesGoogle Scholar
- Colin Allen, Gary Varner and Jason Zinser. 2000. Prolegomena to Any Future Artificial Moral Agent, J. Exp. Theor. Artif. Intell. 12 (July 2000), 251–261. DOI: https://doi.org/10.1080/09528130050111428Google ScholarCross Ref
- Mark Coeckelbergh. 2016. Responsibility and the moral phenomenology of using self-driving cars. Appl. Artif. Intell, 3, 8 (Nov. 2016), 748–57. DOI: https://doi.org/10.1080/08839514.2016.1229759Google Scholar
- Mark Coeckelbergh. 2009. Virtual moral agency, virtual moral responsibility: on the moral significance of the appearance, perception and performance of artificial agents. AI & Soc. 24 (May 2009), 181-189. DOI: https://doi.org/10.1007/s00146-009-0208-3Google Scholar
- John-Stewart Gordon. 2020. Building moral Robots: Ethical pitfalls and challenges. Sci. Eng. Ethics. 26 (Feb. 2020), 141-157. DOI: https://doi.org/10.1007/s11948-019-00084-5Google Scholar
- Sven Nyholm. 2018. Attributing agency to automated systems: Reflections to human-robot collaborations and responsibility-loci. Sci. Eng. Ethics 24 (Aug: 2018), 1201-1219. DOI: https://doi.org/10.1007/s11948-017-9943-xGoogle Scholar
- Rose de Jong. 2019. The retribution-gap and responsibility-loci related to robots and automated technology: A reply to Nyholm. Sci Eng Ethics 26 (Jul. 2019), 727–735. DOI: https://doi.org/10.1007/s11948-019-00120-4Google Scholar
- Luciano Floridi. 2016. Faultless responsibility: on the nature and allocation of moral responsibility for distributed moral actions. Philos. Trans. A. Math Phys. Eng. Sci. 374, 2083 (Dec. 2016), 20160112. DOI: https://doi.org/10.1098/rsta.2016.0112Google Scholar
- Wendy Liu. 2008. Focusing on desirability: the effect of decision interruption and suspension of preferences. J. Consum. Res. 35 (Dec. 2008), 640-652. DOI: https://doi.org/10.1086/592126Google ScholarCross Ref
- Daniel Kahneman. 2003. Perspective on judgment and choice. Am. Psychol. 58, 9 (Sept. 2003), 697–720. DOI: https://doi.org/10.1037/0003-066X.58.9.697Google ScholarCross Ref
- Laura Zwaan, Sandra Monteiro, Jonathan Sherbino, Jonathan Ilgen, Betty Howey and Geoffrey Norman. 2017. Is bias in the eye of the beholder? A vignette study to assess recognition of cognitive biases in clinical case workups. BMJ Qual. Saf. 26, 2 (Feb. 2017), 104-110. DOI: https://doi.org/10.1136/bmjqs-2015-005014Google ScholarCross Ref
- Stefano Guidi, Oronzo Parlangeli, Sandro Bettella and Sergio Roncato. 2011. Features of the selectivity for contrast polarity in contour integration revealed by a novel tilt illusion. Perception 40, 11 (Nov. 1), 1357-1375. DOI: https://doi.org/10.1068/p6897Google ScholarCross Ref
- Oronzo Parlangeli and Sergio Roncato. 2010. Draughtsmen at work. Percept. 39, 2 (Feb. 2010). 255-259. DOI: https://doi.org/10.1068/p6500Google ScholarCross Ref
- Jonathan S. B. T. Evans. 1989. Essays in cognitive psychology. Bias in human reasoning: Causes and consequences (1st. ed.). Lawrence Erlbaum Associates Inc., Hove and London.Google Scholar
- Daniel L. Ames and Susan T. Fiske. 2013. Intentional harms are worse, even when they're not. Psychol Sci 24, 9 (Sept. 2013), 1755–1762. DOI: https://doi.org/10.1177/0956797613480507Google ScholarCross Ref
- Daniel L. Ames and Susan T. Fiske. 2015. Perceived intent motivates people to magnify observed harms. In Proceedings of the National Academy of Sciences, 11, 12 (March 2015), 3599-3605. DOI: https://doi.org/10.1073/pnas.1501592112Google Scholar
- Mark D. Alicke. 1992. Culpable causation. J. Pers. Soc. Psychol. 63, 3 (Sept. 1992), 368–378. DOI: https://doi.org/10.1037/0022-3514.63.3.368Google ScholarCross Ref
- Taemie Kim and Pamela Hinds. 2006. Who should I blame? Effects of autonomy and transparency on attribution in human-robots interaction. In Proceedings of The 15th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN 2006), Hatfiled, UK, 80-85. DOI: https://doi.org/10.1109/ROMAN.2006.314398Google ScholarCross Ref
- Joo-Wha Hong and Dimitri Williams. 2019. Racism, responsibility and autonomy in HCI: Testing perceptions of an AI agent. Comput. Hum. Behav. 100 (June 2019), 79-84. DOI: https://doi.org/10.1016/j.chb.2019.06.012Google Scholar
- Richard D. Johnson, Natasha F. Veltri and Steven Hornik. 2008. Attributions of Responsibility Toward Computing Technology: The Role of Interface Social Cues and User Gender. Int. J. Hum-Comput. Int. 24, 6 (July 2008), 595-612. DOI: https://doi.org/10.1080/10447310802205784Google Scholar
- Daniel B. Shank and Alyssa DeSanti. 2018. Attributions of morality and mind to artificial intelligence after real-world moral violations. Comput. Hum. Behav. 86 (Sept. 2018), 401-411. DOI: https://doi.org/10.1016/j.chb.2018.05.014Google Scholar
- David Premack and Guy Woodruff. 1978. Does the chimpanzee have a theory of mind? The Behavioral and Brain Sciences 1, 4 (Dec. 1978) 515–526. DOI: https://doi.org/10.1017/S0140525X00076512Google ScholarCross Ref
- Daniel C. Dennett. 1987. The intentional stance. MIT press, Cambridge, Ma.Google Scholar
- Nicolas Epley, Adam Waytz and John T. Cacioppo. 2007. On seeing human: A three-factor theory of anthropomorphism. Psychol. Rev. 114, 4 (Nov. 2007), 864-886. DOI: https://doi.org/10.1037/0033-295X.114.4.864Google ScholarCross Ref
- Deborah Kelemen and Susan Carey. 2007. The essence of artifacts: Developing the design stance, in S. Lawrence and E. Margolis (Eds.), Creations of the Mind: Artifacts and their representation, Oxford University Press, Oxford, (2007).Google Scholar
- Kazunori Terada, Takashi Shamoto, Haying Mei and Akira Ito. 2007. Reactive Movements of non-humanoid robots cause intention attribution in humans. In Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems (IEEE/RSJ 2007), October 29 - November 2, 2007, San Diego, CA, USA, 3715-3720. DOI: https://doi.org/10.1109/IROS.2007.4399429Google ScholarCross Ref
- Oronzo Parlangeli, Tommaso Chiantini and Stefano Guidi. 2012. A Mind in a disk: The attribution of mental state to technological systems. Work 41 (2012), 1118-1123. DOI: 10.3233/WOR-2012-0291-1118Google ScholarCross Ref
- Oronzo Parlangeli, Stefano Guidi and Fiore R. Farina. 2012. Overloading Disks onto a Mind: Quantity Effects in the Attribution of Mental States to Technological Systems. Proceedings of the 18th international Congress, IEA – International Ergonomics Association, Recife (2012).Google Scholar
- Oronzo Parlangeli, Maria C. Caratozzolo and Stefano Guidi. 2014. Multitasking and Mentalizing Machines: How the Workload Can Have Influence on the System Comprehension. In D. Harris (ed.) Engineering Psychology and Cognitive Ergonomics. 11th International Conference, EPCE 2014, Held as Part of HCI International 2014, Heraklion, Crete, Greece (2014). DOI: https://doi.org/10.1007/978-3-319-07515-0_6Google Scholar
- Kurt Grey, Liane Young and Adam Waytz. 2012. Mind Perception Is the Essence of Morality. Psychological Inquiry. 23, 101-124. DOI: https://doi.org/10.1080/1047840X.2012.651387Google ScholarCross Ref
- Hideyuki Takahashi, Kazunori Terada, Tomoyo Morita, Shinsuke Suzuki, Tomoki Haji, Hideki Kozima 2014. Different impressions of other agents obtained through social interaction uniquely modulate dorsal and ventral pathway activities in the social human brain. Cortex, 58 (Sept. 2014), 289–300. DOI: https://doi.org/10.1016/j.cortex.2014.03.011Google ScholarCross Ref
- Evgeniya Hristova and Maurice Grinberg. 2015. Should Robots Kill? Moral Judgments for Actions of Artificial Cognitive Agents. In Proceedings of EuroAsianPacific Joint Conference on Cognitive Science (EAPCogSci 2015). CEUR-WS.org., 306-311.Google Scholar
- Bertram F. Malle, Matthias Scheutz, Thomas Arnold, John Voiklis and Corey Cusimano. 2015. Sacrifice one for the good of many?: People apply different moral norms to human and robot agents. In Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction. ACM Press, 117–124. DOI: http://dx.doi.org/10.1145/2696454.2696458Google ScholarDigital Library
- Tetsushi Tanibe, Takaaki Hashimoto and Kaoni Karasawa. 2017. We perceive a mind in a robot when we help it. PLoS ONE 12, 7 (July 2017), 12 pages. DOI: https://doi.org/10.1371/journal.pone.0180952Google ScholarCross Ref
- Heather Grey and Kurt Grey. 2007. Dimensions of Mind Perception. Science. 315, 619 (Feb. 2007). DOI: https://doi.org/10.1126/science.1134475Google ScholarCross Ref
- Carlo Chiorri, Carlo Garofalo and Patrizia Velotti. 2019. Does the Dark Triad manifest Similarly in men and women? Measurement invariance of the Dirty Dozen across sex. Current Psychology, 38, 659-675 (July 2019). DOI: https://doi.org/10.1007/s12144-017-9641-5Google ScholarCross Ref
- Peter K. Jonason, and Gregory D. Webster. 2010. The dirty dozen: A concise measure of the dark triad. Psychological Assessment 22(2), 420-432 (June 2010). DOI: https://doi.org/10.1037/a0019265Google ScholarCross Ref
- Human beings and robots: Are there any differences in the attribution of punishments for the same crimes?
Recommendations
People May Punish, But Not Blame Robots
CHI '21: Proceedings of the 2021 CHI Conference on Human Factors in Computing SystemsAs robots may take a greater part in our moral decision-making processes, whether people hold them accountable for moral harm becomes critical to explore. Blame and punishment signify moral accountability, often involving emotions. We quantitatively ...
On the moral responsibility of military robots
This article discusses mechanisms and principles for assignment of moral responsibility to intelligent robots, with special focus on military robots. We introduce the concept autonomous power as a new concept, and use it to identify the type of robots ...
Robot minds and human ethics: the need for a comprehensive model of moral decision making
Building artificial moral agents (AMAs) underscores the fragmentary character of presently available models of human ethical behavior. It is a distinctly different enterprise from either the attempt by moral philosophers to illuminate the "ought" of ...
Comments