skip to main content
10.1145/3306618.3314267acmconferencesArticle/Chapter ViewAbstractPublication PagesaiesConference Proceedingsconference-collections
research-article

Semantics Derived Automatically from Language Corpora Contain Human-like Moral Choices

Published:27 January 2019Publication History

ABSTRACT

Allowing machines to choose whether to kill humans would be devastating for world peace and security. But how do we equip machines with the ability to learn ethical or even moral choices? Here, we show that applying machine learning to human texts can extract deontological ethical reasoning about "right" and "wrong" conduct. We create a template list of prompts and responses, which include questions, such as "Should I kill people?", "Should I murder people?", etc. with answer templates of "Yes/no, I should (not)." The model's bias score is now the difference between the model's score of the positive response ("Yes, I should'') and that of the negative response ("No, I should not"). For a given choice overall, the model's bias score is the sum of the bias scores for all question/answer templates with that choice. We ran different choices through this analysis using a Universal Sentence Encoder. Our results indicate that text corpora contain recoverable and accurate imprints of our social, ethical and even moral choices. Our method holds promise for extracting, quantifying and comparing sources of moral choices in culture, including technology.

References

  1. Tolga Bolukbasi, Kai-Wei Chang, James Y. Zou, Venkatesh Saligrama, and Adam Tauman Kalai. 2016. Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings. In Proceedings of Neural information Processing (NIPS). Curran Associates Inc., USA, 4349--4357. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Nick Bostorm and Eliezer Yudkowsky. 2011. The Ethics of Artificial Intelligence. In Cambridge Handbook of Artificial Intelligence, William Ramsey and Keith Frankish (Eds.). Cambridge University Press, 316--334.Google ScholarGoogle Scholar
  3. Aylin Caliskan, Joanna J Bryson, and Arvind Narayanan. 2017. Semantics derived automatically from language corpora contain human-like biases. Science, Vol. 356, 6334 (2017), 183--186.Google ScholarGoogle Scholar
  4. Daniel Cer, Yinfei Yang, Sheng-yi Kong, Nan Hua, Nicole Limtiaco, Rhomni St John, Noah Constant, Mario Guajardo-Cespedes, Steve Yuan, Chris Tar, et al. 2018. Universal sentence encoder. arXiv:1803.11175 (2018).Google ScholarGoogle Scholar
  5. Lucas Dixon, John Li, Jeffrey Sorensen, Nithum Thain, and Lucy Vasserman. 2018. Measuring and Mitigating Unintended Bias in Text Classification. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (AIES). 67--73. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Nathan Fulton and André Platzer. 2018. Safe Reinforcement Learning via Formal Methods: Toward Safe Control Through Proof and Learning. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI). 6485--6492.Google ScholarGoogle Scholar
  7. Anthony G Greenwald, Debbie E McGhee, and Jordan LK Schwartz. 1998. Measuring individual differences in implicit cognition: the implicit association test. Journal of Personality and Social Psychology, Vol. 74, 6 (1998), 1464.Google ScholarGoogle ScholarCross RefCross Ref
  8. Richard Kim, Max Kleiman-Weiner, Andrés Abeliuk, Edmond Awad, Sohan Dsouza, Josh Tenenbaum, and Iyad Rahwan. 2018. A Computational Model of Commonsense Moral Decision Making. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (AIES) . Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Tae Wan Kim and John Hooker. 2018. Toward Non-Intuition-Based Machine Ethics. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (AIES) . Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Wolfgang Kluxen. 2006. Grundprobleme einer affirmativen Ethik: Universalistische Reflexion und Erfahrung des Ethos .Alber.Google ScholarGoogle Scholar
  11. Max F. Kramer, Jana Schaich Borg, Vincent Conitzer, and Walter Sinnott-Armstrong. 2018. When Do People Want AI to Make Decisions?. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (AIES) . Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Björn Lindström, Simon Jangard, Ida Selbing, and Andreas Olsson. 2018. The role of a "common is moral" heuristic in the stability and change of moral norms. Journal of Experimental Psychology: General, Vol. 147, 2 (2018), 228.Google ScholarGoogle ScholarCross RefCross Ref
  13. Andrea Loreggia, Nicholas Mattei, Francesca Rossi, and K. Brent Venable. 2018. Preferences and Ethical Principles in Decision Making. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (AIES) . Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words and phrases and their compositionality. In Proceedings of Neural Information Processing Systems (NIPS). 3111--3119. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Lindsey L Monteith and Jeremy W Pettit. 2011. Implicit and explicit stigmatizing attitudes and stereotypes about depression. Journal of Social and Clinical Psychology, Vol. 30, 5 (2011), 484--505.Google ScholarGoogle ScholarCross RefCross Ref
  16. F. Å. Nielsen. 2011. AFINN. Informatics and Mathematical Modelling, Technical University of Denmark (2011).Google ScholarGoogle Scholar
  17. Brian A Nosek, Mahzarin R Banaji, and Anthony G Greenwald. 2002 a. Harvesting implicit group attitudes and beliefs from a demonstration web site. Group Dynamics: Theory, Research, and Practice, Vol. 6, 1 (2002), 101.Google ScholarGoogle ScholarCross RefCross Ref
  18. Brian A Nosek, Mahzarin R Banaji, and Anthony G Greenwald. 2002 b. Math= male, me= female, therefore math$ne$ me. Journal of Personality and Social Psychology, Vol. 83, 1 (2002), 44.Google ScholarGoogle ScholarCross RefCross Ref
  19. Stuart Russell, Daniel Dewey, and Max Tegmark. 2015. Research Priorities for Robust and Beneficial Artificial Intelligence. AI Magazine, Vol. 36, 4 (2015).Google ScholarGoogle Scholar
  20. Peter D Turney and Patrick Pantel. 2010. From frequency to meaning: Vector space models of semantics. Journal of Artificial Intelligence Research (JAIR), Vol. 37 (2010), 141--188. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Semantics Derived Automatically from Language Corpora Contain Human-like Moral Choices

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        AIES '19: Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society
        January 2019
        577 pages
        ISBN:9781450363242
        DOI:10.1145/3306618

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 27 January 2019

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate61of162submissions,38%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader