Skip to main content
Log in

Learning robots interacting with humans: from epistemic risk to responsibility

  • Original Article
  • Published:
AI & SOCIETY Aims and scope Submit manuscript

Abstract

The import of computational learning theories and techniques on the ethics of human-robot interaction is explored in the context of recent developments of personal robotics. An epistemological reflection enables one to isolate a variety of background hypotheses that are needed to achieve successful learning from experience in autonomous personal robots. The conjectural character of these background hypotheses brings out theoretical and practical limitations in our ability to predict and control the behaviour of learning robots in their interactions with humans. Responsibility ascription problems, which concern damages caused by learning robot actions, are analyzed in the light of these epistemic limitations. Finally, a broad framework is outlined for ethically motivated scientific inquiries, which aim at improving our capability to understand, anticipate, and selectively cope with harmful errors by learning robots.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1

Similar content being viewed by others

Notes

  1. Failures of this safety policy in industrial environments are witnessed by a significant number of accidents involving robots in factories and plants. Useful information about robots and safety of human beings is provided in the 8 Jun 2006 issue of The Economist-Technology Quarterly. An electronic copy of the article is available on line at: http://www.economist.com/displaystory.cfm?story_id = 7001829.

  2. The overall rule governing the behaviour of an unsociable robot is relatively easy to state, but its actual implementation raises non-trivial theoretical and technological problems, which include the need for real-time reactivity and motion planning in high-dimensional configuration spaces. Furthermore, the reliability of the proposed solutions usually declines sharply when the environment becomes more and more cluttered, dense, and complex. A survey of effective methods and solutions to such problems can be found in Minguez and Montano (2004); Brock and Khatib (2002); and Kohout (2000).

  3. This sweeping claim is clearly stated and motivated in Cucker and Smale (2001). Mitchell (1997) (p. 42ff.) is also a valid source for a discussion of inductive biases needed by computational learning agents.

  4. Several proposals for relaxing some of these constraints have been advanced, but the modified learning algorithms are usually intractable for all but the smallest problems. Even though a few algorithms (see, e.g., Pineau and Gordon 2005) solved some such computational problems and demonstrated competitive performances in limited tasks, their use is still far from being widespread in robotics.

  5. Background conjectural assumptions about the environment play a crucial role here too, insofar as successful application of RL learning depends on the correctness of these assumptions about the environment. A detailed argument to this effect is provided, in connection with RL algorithms for adapting navigation control strategies in behaviour-based robotic architectures, in Datteri et al. (2006).

  6. The first influential work is due to Vapnik and Chervonenckis (1971), while a comprehensive overview of the resulting theory (known as VC theory or statistical learning theory) was provided by Vapnik (1999a, b).

  7. A solution for a class of learning problems is the specification of a learning algorithm that can be trained by means of a suitable set of training examples.

  8. From the standpoint of computational complexity theory this is usually taken to mean that the learning problem belongs to the class P.

  9. The connections between PAC models and (the theory of) empirical processes were first exploited by Blumer and colleagues (1989); thereafter, many efforts have been produced to achieve a better understanding of these connections (see, e.g., Vidyasagar 1996).

  10. Clearly, in this toy example, computational aspects are difficult to appreciate; however, these aspects are crucial in the real case. The more the parameters are the more expensive it is to explore the parameter space in order to find the “best” solution. Of course, a line can be represented by only two parameters while a curve requires (in general) more parameters.

  11. For discussion, see Tamburrini (2006); for an analysis of early cybernetic reflections on the use of learning machines, see Cordeschi and Tamburrini (2005).

  12. This error is referred to as false negative.

  13. More details about the effectiveness of ROC curves are found in Zweig and Campbell (1993), while some practical issues are more extensively discussed in Fawcett (2004).

References

  • Bishop CM (1995) Neural networks for pattern recognition. Oxford University Press, Oxford

    Google Scholar 

  • Blumer A, Ehrenfeucht A, Haussler D, Warmuth MK (1989) Learnability and the Vapnik-Chervonenkis dimension. J ACM 36(4):929–965

    Article  MathSciNet  MATH  Google Scholar 

  • Brock O, Khatib O (2002) Elastic strips: a framework for motion generation in human environments. Int J Robot Res 21(12):1031–1052

    Article  Google Scholar 

  • Capurro R, Nagenborg M, Weber J, Pingel C (2006) Methodological issues in the ethics of human-robot interaction. In: Tamburrini G, Datteri E (eds) Ethics of human interaction with robotic, bionic, and AI systems, workshop book of abstracts. Istituto Italiano per gli Studi Filosofici, Napoli, p 9

    Google Scholar 

  • Cordeschi R, Tamburrini G (2005) Intelligent machinery and warfare: historical debates and epistemologically motivated concerns. In: Magnani L, Dossena R (eds) Computing, philosophy, and cognition. King’s College Publications, London, pp 1–23

    Google Scholar 

  • Cucker F, Smale S (2001) On the mathematical foundations of learning. Bull Am Math Soc 39(1):1–49

    Article  MathSciNet  Google Scholar 

  • Datteri E, Hosni H, Tamburrini G (2006) An inductionless and default-based analysis of machine learning procedures. In: Magnani L (eds) Model based reasoning in science and engineering. College Publications, London, pp 379–399

    Google Scholar 

  • Fawcett T (2004) ROC graphs: notes and practical considerations for researchers. Tech report HPL-2003–4, HP Laboratories

  • Haussler D (1990) Probably approximately correct learning. In: AAAI-90 proceedings of the eight national conference on artificial intelligence, 1101–1108

  • Kohout R (2000) Challenges in real-time obstacle avoidance. In: AAAI spring symposium on real-time autonomous systems. Palo Alto

  • Matthias A (2004) The responsibility gap: ascribing responsibility for the actions of learning automata. Ethics Inf Technol 6:175–183

    Article  Google Scholar 

  • Minguez J, Montano L (2004) Nearness diagram navigation (ND): collision avoidance in troublesome scenarios. IEEE Trans Rob Autom 20(1):45–59

    Article  Google Scholar 

  • Mitchell TM (1997) Machine learning. McGraw Hill New York

    MATH  Google Scholar 

  • Müller KR, Mika S, Ratsch G, Tsuda K, Scholkopf B (2001) An introduction to kernel-based learning algorithms. IEEE Trans Neural Netw 12(2):181–201

    Article  Google Scholar 

  • Pineau J, Gordon G (2005) “POMDP planning for robust robot control”. International Symposium on Robotics Research (ISRR), San Francisco

  • Pitt L, Valiant L (1988) Computational limitations on learning from examples. J ACM 35:965–984

    Article  MathSciNet  MATH  Google Scholar 

  • Tamburrini G (2006) AI and Popper’s solution to the problem of induction. In: Jarvie I, Milford K, Miller D (eds) Karl popper: a centennial assessment, metaphysics and epistemology. Ashgate, London, 2:265–282

  • Valiant L (1984) A theory of the learnable. Commun ACM 27:1134–1142

    Article  MATH  Google Scholar 

  • Vapnik VN (1999a) An overview of statistical learning theory. IEEE Trans Neural Netw 10(5)

  • Vapnik VN (1999b) The nature of statistical learning theory. Springer, Heidelberg

    Google Scholar 

  • Vapnik VN, Chervonenkis AY (1971) On the Uniform Convergence of relative frequencies of events to their probabilities. Theor Probab Appl 16:264–280

    Article  MATH  Google Scholar 

  • Vidyasagar M (1996) A theory of learning and generalization. Springer, Heidelberg

    Google Scholar 

  • Zweig MH, Campbell G (1993) Receiver-operating characteristic (ROC) plots: a fundamental evaluation tool in clinical medicine. Clin Chem 39(4):561–577

    Google Scholar 

Download references

Acknowledgements

The financial support of the VI Framework Programme of the European Union (Science and Society, contract number 017759) is gratefully acknowledged.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guglielmo Tamburrini.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Santoro, M., Marino, D. & Tamburrini, G. Learning robots interacting with humans: from epistemic risk to responsibility. AI & Soc 22, 301–314 (2008). https://doi.org/10.1007/s00146-007-0155-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00146-007-0155-9

Keywords

Navigation