Elsevier

Biosystems

Volume 128, February 2015, Pages 37-47
Biosystems

Inference of other’s internal neural models from active observation

https://doi.org/10.1016/j.biosystems.2015.01.005Get rights and content

Abstract

Recently, there have been several attempts to replicate theory of mind, which explains how humans infer the mental states of other people using multiple sensory input, with artificial systems. One example of this is a robot that observes the behavior of other artificial systems and infers their internal models, mapping sensory inputs to the actuator’s control signals. In this paper, we present the internal model as an artificial neural network, similar to biological systems. During inference, an observer can use an active incremental learning algorithm to guess an actor’s internal neural model. This could significantly reduce the effort needed to guess other people’s internal models. We apply an algorithm to the actor–observer robot scenarios with/without prior knowledge of the internal models. To validate our approach, we use a physics-based simulator with virtual robots. A series of experiments reveal that the observer robot can construct an “other’s self-model”, validating the possibility that a neural-based approach can be used as a platform for learning cognitive functions.

Introduction

Robots can represent a simplified model of human behavior, whereby the robot senses its environment and reacts to various input signals. The robot’s ‘brain’ controls its body in response to the input signals using artificial neural networks. The topology and weights of the neural network characterize the behavioral properties of the robot. Recently, several investigations have used robots in order to gain insight into human cognition by creating a simplified analogous problem (Bongard et al., 2006, Webb, 2001, Floreano and Keller, 2010). Bongard et al. built a starfish robot; however, it was unaware of its own body shape (Bongard et al., 2006). Using an estimation–exploration algorithm (EEA) (Bongard and Lipson, 2007), the robot was able to successfully create a self-model of its body shape using an iterative estimation and exploration procedure. In the estimation step, the robot searched multiple candidates to determine its body shape. Subsequently, in the exploration step, the algorithm determined the actions that most strongly agreed with the multiple candidate body shapes.

Unlike self-modeling, however, theory of mind (ToM) is a high-level cognitive function that models the mental states (beliefs, intents, desires, imagination, knowledge, etc.) of another entity. In robotic studies, robots have demonstrated the ability to mimic the behavior of humans or to decode the intentions of a third party (both human and robot). For example, Scassellati implemented Baron-Cohen’s ToM model for the humanoid robot COG (Scassellati, 2002). Breazeal et al. demonstrated that an animal-like robot could pass the false-belief test widely used to test ToM in young children (Breazeal et al., 2005). Furthermore, Buchsbaum et al. carried out simulations in which one agent attempted to determine another agent’s behavior using rat-like characters (Buchsbaum et al., 2005). In this particular study, the observer exploited his own behavior tree to infer others’ intentions.

However, few reports have described the representation of another entity’s mind as a neural circuit. Revealing an internal neural model based on observations is a challenging task. However, there is great potential for using neural networks as internal models, because it would mimic the underlying mechanisms of human representations in the form of neural connections. Many different definitions of the self and other’s self-representations exist, ranging from symbolic states to complex neural models. For example, Bongard et al. (Bongard et al., 2006) used the morphological structure of a robot as a self-model. The robot had no physical model of itself on which to base an understanding, and attempted to construct models of its body using iterative estimation–exploration steps. Kim and Lipson used a simple feed-forward network to represent the minds of other (Kim and Lipson, 2009a, Kim and Lipson, 2009b, Kim and Lipson, 2009b).

In this paper, we propose the use of active incremental learning to infer the internal neural models of other entities both with and without prior knowledge (Fig. 1). We used two robots, referred to as the actor and the observer. The actor used a neural controller (implemented as an artificial neural network) to control its behavior based on sensory information. The observer monitored the behaviors of the actor and attempted to infer the actor’s internal model from these observations. The observer used the inferred self-model of the actor to predict the actor’s future behavior. In this approach, instead of programming the other’s internal model manually, the observer attempted to predict the other’s self-model interactively. The observer robot started from a single actor trajectory and invited the actor robot to demonstrate additional trajectories, which were then used to infer information about the actor’s self-model using the EEA method (Bongard and Lipson, 2007).

In particular, we tested the impact that prior knowledge had on the actor’s internal model. Initially, we assumed that the actor and observer were the same species and that the observer could use his self-model (neural topology). Therefore, the ToM problem is formulated as the inference of the connection weights given the shared structure. We subsequently assumed that the two robots are different species and that the actor could not use his self-model for the ToM. As a result, the observer needs to search for the architecture of the neural network and the weights simultaneously to infer the other’s self-model. We used a physics-based simulation to run the ToM experiments, which show the potential of this approach given the two experimental conditions.

The rest of this paper is organized as follows. In Section 2 we describe related research, including the research on ToM in robots. In Section 3 we apply the estimation–exploration algorithm for the robotic ToM. Finally, in Section 4, we present our experimental results.

Section snippets

Inference of other’s mind in humans

ToM is the ability to attribute mental states to oneself and others, and to understand that others have different beliefs, desires, and intentions from one’s own (Premack and Woodruff, 1978). The first paper on ToM, published in 1978 by Premack and Woodruff, posed the question, “Does the chimpanzee have a theory of mind”? Since then, many articles on ToM in human and non-human primates have been published (Call and Tomasello, 2008). Attempts have been made to reveal the existence of ToM in many

Proposed method

In this paper, we propose the use of active incremental learning to infer an observer’s internal model. There are two robots in this environment; one robot is an actor and the other is an observer. The actor controls itself using a feed-forward neural network (NN), with the inputs to the NN being sensory information, and the outputs control the speed and direction of the robot. The observer monitors the behavior of the actor and attempts to infer its internal neural model with/without prior

Experimental results and discussion

In this research, we performed experiments using a virtual robot in physics-based simulation (PhysX) environments. The results were averaged over 10 runs.

Concluding remarks

We used a reverse-engineering algorithm to construct a model of the internal neural network of robots using observations of their behavior. The observer robots actively collected information on the actor’s trajectory toward a goal and inferred an internal model based on the behavior. A series of experiments showed that the proposed method can be useful in identifying internal models. Furthermore, this research demonstrated the possibility of using a neural-based approach as a platform for

Acknowledgements

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korean Government (MSIP) (2013 R1A2A2A01016589, 2010-0018948, 2010-0018950). The authors would like to express thanks to Prof. Hod Lipson for his guidance on the early version of this work.

References (35)

  • J. Call et al.

    Does the chimpanzee have a theory of mind? 30 years later

    Trends Cogn. Sci.

    (2008)
  • K.-J. Kim et al.

    Automated synthesis of multiple analog circuits using evolutionary computation for redundancy-based fault-tolerance

    Appl. Soft Comput.

    (2012)
  • R. Saxe

    Theory of mind (neural basis)

    Encyclopedia Consciousness

    (2009)
  • S. Baron-Cohen

    Mind Blindness: An Essay on Autism and Theory of Mind

    (1995)
  • J. Bongard et al.

    Automated reverse engineering of nonlinear dynamical systems

    Proc. NatL. Acad. Sci.

    (2007)
  • J. Bongard et al.

    Resilient machines through continuous self-modeling

    Science

    (2006)
  • T. Bosse et al.

    A two-level BDI-agent model for theory of mind and its use in social manipulation

    Proceedings of the Artificial and Ambient Intelligence Conference

    (2007)
  • C. Breazeal et al.

    Learning from and about others: Towards using imitation to bootstrap the social understanding of others by robots

    Artif. Life

    (2005)
  • A. Bringsjord et al.

    Toward logic-based cognitively robust synthetic characters in digital environments

    Proceedings of the First Artificial General Intelligence

    (2008)
  • D. Buchsbaum et al.

    A simulation-theory inspired social learning system for interactive characters

    IEEE International Workshop on Robots and Human Interactive Communication

    (2005)
  • Y. Demiris et al.

    Distributed, predictive perception of actions: a biologically inspired robotics architecture for imitation and learning

    Connect. Sci.

    (2003)
  • D. Floreano et al.

    Evolution of adaptive behavior in robots by means of Darwinian selection

    PLOS Biol.

    (2010)
  • Y. Freund et al.

    Selective sampling using the query by committee algorithm

    Mach. Learn.

    (1997)
  • E. Herrmann et al.

    Humans have evolved specialized skills of social cognition: the cultural intelligence hypothesis

    Science

    (2007)
  • R.E. Kaliouby et al.

    Mind reading machines: automated inference of cognitive mental states from video

    IEEE International Conference on Systems, Man and Cybernetics

    (2004)
  • R. Kelley et al.

    An architecture for understanding intent using a novel hidden markov formulation

    Int. J. Hum. Robot.

    (2008)
  • K.-J. Kim et al.

    Theory of mind in simulated robots

    Genetic and Evolutionary Computation Conference (GECCO) – Late Braking Papers

    (2009)
  • Cited by (9)

    • Calibration and extended validation of a virtual asphalt mixture compaction model using bullet physics engine

      2021, Construction and Building Materials
      Citation Excerpt :

      Hence, the main focus of physics engines was the speed of the simulation to provide a realistic gaming experience for the video game players. Subsequently, with an increase in accuracy of the physics engines, they were used in various areas of research including medical robotics [29,30], psychology [31], petroleum engineering [32], agricultural machinery safety [33], earthquake engineering [34–36], manufacturing systems [37], granular flow [38], disaster management [39] and geotechnical engineering [40–45]. Physics engines typically use a simplified contact model and use both the central processing unit (CPU) and graphics processing unit (GPU) which in turn makes them faster as compared to DEM.

    • Feasibility of using a physics engine to virtually compact asphalt mixtures in a gyratory compactor

      2021, Construction and Building Materials
      Citation Excerpt :

      Physics engines were originally developed for computer games to produce a realistic gaming experience. Subsequently, they have also been used in various areas of research including medical robotics [55,56], psychology [57], petroleum engineering [58], agricultural machinery safety [59], earthquake engineering [60–62], manufacturing systems [63], granular flow [64] and disaster management [65]. They are typically faster [66] than DEM because of the use of (i) a simplified contact model and (ii) both the central processing unit (CPU) and graphics processing unit (GPU) in the simulation.

    • Physics engine based simulation of shear behavior of granular soils using hard and soft contact models

      2021, Journal of Computational Science
      Citation Excerpt :

      In recent years, driven by the rapid development and high competitiveness of the computer gaming and movie industry, the accuracy, computational speed, and functionalities of physics engine techniques have been significantly improved. Today, physics engines are increasingly used as scientific computational platform in various disciplines, including robotic control [5,6], crowd simulation [7], biomedical engineering [8,9], autonomous vehicle research [10], virtual and augmented reality [11], and psychological research [12]. The granular soil particles can be simulated as rigid bodies, if the particle breakage is not considered.

    View all citing articles on Scopus
    View full text