Evolution of behaviors in autonomous robot using artificial neural network and genetic algorithm

https://doi.org/10.1016/S0020-0255(03)00125-7Get rights and content

Abstract

In distributed autonomous robot (agents) systems, each robot (predator or prey) must behave by itself according to its states and environments, and if necessary, must cooperate with other robots in order to carry out a given task. Therefore it is essential that each robot have both learning and evolution ability to adapt to dynamic environment. This paper proposes a pursuing system utilizing the artificial life concept where autonomous mobile robots emulate social behaviors of animals and insects and realize their group behaviors. Each robot contains sensors to perceive other robots in several directions and decides its behavior based on the information obtained by the sensors. In this paper, a neural network is used for behavior decision controller. The input of the neural network is decided by the existence of other robots and the distance to the other robots. The output determines the directions in which the robot moves. The connection weight values of this neural network are encoded as genes, and the fitness individuals are determined using a genetic algorithm. Here, the fitness values imply how much group behaviors fit adequately to the goal and can express group behaviors. The validity of the system is verified through simulation. Besides, in this paper, we could have observed the robots’ emergent behaviors during simulation.

Introduction

When facing difficult problems, we can often find the solution from creatures in nature. From this kind of attempts, an research area of artificial life (AL) came into being by Langton in 1987, in purpose of unifying hitherto individually achieved research results of living creatures and further activating those researches [1], [2], [3], [4].

In engineering aspects, the goal of the artificial life is to incarnate the unique behaviors or phenomena of living creatures in nature onto artifacts like computers. We expect the artificial life can provide a useful methodology for the autonomous mobile robot learning which is full of autonomy and creativity. One of basic concept of the artificial life is “emergence”. According to its mechanism, the emergence can be divided into several classes. Most of them determine behaviors and interaction rules of lower-level components. The local interaction between components generates global orders or behaviors [5], [6], [7], [8], [9], [10], [11], [12], [13].

The present paper develops a pursuit system that is mostly applied to the area of robot and robot in order to apply an emergent characteristic of the artificial life into machine learning. The pursuit system contains robots and preys. In the system, robots catch the preys escaping from them through evolution. For the simulator of the system, we have modeled the structure of the robot with the artificial neural network and evolved the structure of the neural network with the genetic algorithm. The prey is given only pre-knowledge of direction identification to move.

Numerous researches about the autonomous mobile robot control in the pursuit system have been studied. Nolfi and Foreano [23] simulated pursuit system with two robots (a predator and a prey) in real environments. They evolved both robots reciprocally with genetic algorithm. Yasuo and Shin [15] controlled the autonomous mobile robots using reinforcement learning. Kam-Chuen [18] evolved the autonomous mobile robots with the genetic algorithm. By II-Kwon [19], both fuzzy controller and genetic algorithm were used. The fuzzy function displayed robot’s position, and the genetic algorithm was used for learning. The reinforcement learning method is to develop robot’s behaviors by means of the interrelationship with environment and resulting reinforcement signals. It can guarantee learning and adaptability without precise pre-knowledge about environments. However, its critical weak point is the difficulty of learning when the rewards of taken actions are not instantly computed. Takayuki and Yasuo [14], [15] have a problem in applying to the open environments which is dynamically changing since once a robot’s role is decided, then it is fixed. By Jeong and Lee [19] and Pena-Reyes and Sipper [20] both robots and prey move randomly, and the genetic algorithm helps to prevent the collision between robots. However, robots were not intelligent to do learning for themselves. Also, since the gene structure was represented by fuzzy member functions, it took much time to resolve pursuing problems.

Therefore, in this paper, to resolve those problems, we apply the artificial life method in our system. We model the robot structure with the artificial neural network that is easy to model both robots and use the genetic algorithm for robot learning, which results in giving intelligence to robots in their pursuing movement toward the prey. The prey has only pre-knowledge to identify distance to avoid the robots. We have designed a virtual environment where 20 robots and a prey. Furthermore, various selection methods of the genetic algorithm are considered for simulation.

The rest of this paper is organized as follows. In Section 2, the related works of this paper and is described. Section 3 describes the evolution of autonomous robots in pursuit system. In Section 4, simulation and evaluation are described. Finally, conclusion is presented in Section 5.

Section snippets

Genetic algorithm

A genetic algorithm (GA) is a method to obtain an optimal solution by applying a theory of biological evolution [21], [22]. GAs generally consist of three fundamental operators: reproduction, crossover and mutation. Given an optimization problem, simple GAs encode the concerned parameters into finite bit strings, and then run iteratively using the three operators in a random way but based on the fitness function evolution to perform the basic tasks of copying strings, exchanging portions of

Virtual environment and the structure of the system

Virtual environment of the pursuit system is a 10 × 10 lattice where up-and-down and right-and-left are connected each other as shown and robots and preys coexist, as shown in Fig. 2. Accordingly, if a robot or a prey move continuously toward one direction, it comes back to its origin. One any spot, there exists only one robot or one prey. In this paper, the prey is given the pre-knowledge that can identify and avoid robots in four directions. If it cannot move any more, it can stay on a spot.

The

Simulation

This section describes emergent behaviors of autonomous mobile robots in pursuit system. The parameters used for simulation are described in Table 3.

The prey is given a primitive priority in moving directions as left, right, up, and down with decreasing order of priority. Since robots are prey are supposed to be placed on different spots (one on one) on the lattice, there will not occur any collision among them. Each robot continuously evolves during its life time (20 movements), regardless of

Conclusion

This paper merges the artificial life method in emergent behavior evolution of autonomous mobile robot as modeling the pursuit system with the artificial neural network and genetic algorithm. In doing so, in virtual environments, we construct a neural network representing robot and prey and propose a model to evolve its structure. Also, we have compared evolution strategies of several selection methods varying the crossover and mutation rates in simulation and investigated the variation of

References (25)

  • T. Haynes et al.

    Learning cases to resolve conflicts and improve group behavior

    International Journal of Human Computer Studies

    (1998)
  • D. Floreano

    Evolutionary robotics in artificial life and behavior engineering

  • F. Gomez et al.

    Incremental evolution of complex general behavior

    Adaptive Behavior

    (1997)
  • D. Floreano et al.

    Co-evolution and ontogenetic change in competing robots

  • D.W. Lee, K.B. Sim, Behavior learning and evolution of collective autonomous mobile robots using distributed genetic...
  • D.W. Lee, K.B. Sim, Development of communication system for cooperative behavior in collective autonomous mobile...
  • D.W. Lee, H.B. Jun, K.B. Sim, Artificial immune system for realization of cooperative strategies and group behavior in...
  • K.B Sim

    Realisation of intelligent robot system based on artificial life

    Journal of Korean Electronics

    (1997)
  • L.M.J. Hogg et al.

    Socially intelligent reasoning for autonomous robots

    IEEE Transactions on Systems, Man, and Cybernetics––Part A: Systems and Humans

    (2001)
  • M. Dorigo et al.

    Ant system: optimization by a colony of cooperating robots

    IEEE Transactions on Systems, Man, and Cybernetics––Part B: Cybernetics

    (1996)
  • Cited by (37)

    • Navigation of multiple mobile robots in a highly clutter terrains using adaptive neuro-fuzzy inference system

      2015, Robotics and Autonomous Systems
      Citation Excerpt :

      They implemented their technique in Khepera II mobile robot. Lee [8] has used a neural network for behaviour decision controller. The input of the neural network was decided by the distance and existence of other robots.

    • Dual hierarchical genetic-optimal control: A new global optimal path planning method for robots

      2014, Journal of Manufacturing Systems
      Citation Excerpt :

      The optimal path is found by generating random points all over the possible search area, comparing value of cost function in these points and then selecting the best set of points. Random search techniques like annealing simulated, genetic algorithm and neural network methods [1,2] are in this category. In the second group, the analytical gradients of cost function are used for the optimization process.

    • Behavioral modeling with the new bio-inspired coordination generalized molecule model algorithm

      2013, Information Sciences
      Citation Excerpt :

      The notion of goal-oriented behavior is based on the operational notion of goal and “purposive behavior” as proposed by Rosenblueth and Wiener [18], and has been further developed, in psychology, by Miller, Galanter and Pribram [15]. Here is a quick summary of the related works on social behavior: (1) Malrey Lee translates the artificial life method in emergent behavior evolution of autonomous mobile robots into a modeling of the pursuit system with artificial neural networks and genetic algorithms [12]. ( 2) In [9], Hervé Frezza-Buet and Frédéric Alexandre present a cortical model, which is inspired by biological data, to allow an animat to perceive and act on its environment to learn the regularities of the environment and to use them to satisfy certain goals; they then describe this model at different levels of abstraction, from both a biological and a computational point of view. (

    • Environmental framework to visualize emergent artificial forest ecosystems

      2013, Information Sciences
      Citation Excerpt :

      Ecosystem simulation is also used by computer graphics to render a 3D ecosystem scene into an image [9,63]. To, generally, be able to simulate life, artificial life models exist, such as [38,22,27,17]. In simulating life of trees within a forest ecosystem, the spontaneous development [46,23] is one of the main processes of forest development which must be modeled.

    View all citing articles on Scopus

    This work was supported by the Post-doctoral Fellowship Program of Korea Science & Engineering Foundation (KOSEF).

    View full text