Topological navigation and qualitative localization for indoor environment using multi-sensory perception

https://doi.org/10.1016/S0921-8890(02)00276-2Get rights and content

Abstract

This article describes a navigation system for a mobile robot which must execute motions in a building; the robot is equipped with a belt of ultrasonic sensors and with a camera. The environment is represented by a topological model based on a Generalized Voronoi Graph (GVG) and by a set of visual landmarks. Typically, the topological graph describes the free space in which the robot must navigate; a node is associated to an intersection between corridors, or to a crossing towards another topological area (an open space: rooms, hallways, …); an edge corresponds to a corridor or to a path in an open space. Landmarks correspond to static, rectangular and planar objects (e.g. doors, windows, posters, …) located on the walls. The landmarks are only located with respect to the topological graph: some of them are associated to nodes, other to edges. The paper is focused on the preliminary exploration task, i.e. the incremental construction of the topological model. The navigation task is based on this model: the robot self-localization is only expressed with respect to the graph.

Introduction

Navigation is a critical task for a mobile robot to allow it to move and act autonomously in its environment. Because internal sensors on the robot are not accurate enough or may give false measurements, a navigation system must be based on exteroceptive sensors like cameras, sonars or laser range finders.

As opposed to the classical methods based on explicit localization of the robot with respect to the environment, other methods [6], [9] make the robot localization relative only to discriminant features learned and successively perceived by the robot or relative to an area (environment modeling in topologically independent areas: corridors, room, …) [1]; the continuity of a path is guaranteed by a graph which expresses some relationships between landmarks; for example, landmark A is connected to landmark B only if B is visible from A, or if a sensor-based motion (wall following, visual servoing, for example) can be executed to go from A to B. This kind of approach could be more generally embedded in the family of qualitative or topological navigation methods. Note that these methods alone will never be sufficient to provide a truly reliable navigation system in a general indoor environment but have to be integrated in a more general, adaptive system as the ones described in [5].

This paper proposes such a topological navigation ability. A service robot must execute motions in an office environment, so the Generalized Voronoi Graph (GVG) representation proposed by Choset and coworkers [2], [7] could be well adapted to solve the navigation problem; in such a graph, nodes are associated to transitions between areas (corridor crossings, area entrances, doors, …); an edge typically corresponds to a path in a corridor or in an open space. In a corridor, the robot motion can be controlled using sonars to maintain the robot on the GVG; the robot localization is expressed according to the GVG (the robot is on this node or is moving on this edge).

Nevertheless, a self-localization problem may occur because this kind of environment is very ambiguous using only sonars whenever human presence or topological modification (an open door) may occur. If the robot is equipped with several sensors—in our experiment, monocular vision and sonars—it can take advantage of different topological representations (visual landmarks and GVG) to validate an hypothesis about a node recognition. Vision gives stable, reliable information from a large part of the environment, which may be helpful in comparison with ultrasonic sensors.

Kortenkamp and Weymouth [4] have already presented results combining these two sensors. Predetermined forms of gateways are searched with sonar and these distinct places are associated with some simple visual landmarks. The learning phase was not done autonomously, and the processing steps of sonar data made the algorithm usable in only orthogonal corridors. Our contribution is two-fold:

  • we make the robot learn autonomously the environment model, without preliminary guided route traversals and with extended environment structural configurations, although considering only corridors-based environment;

  • we use a visual landmark intrinsic representation independent from the viewpoint and as stable as possible with respect to illumination, scale changes and small occlusions.

Section 2 proposes an overview on the environment representation and the navigation system. 3 Learning the topological graph, 4 Learning landmarks present our strategy to build a hybrid topological map—landmark-based and GVG; in Section 5, experimental results are commented. Finally in Section 6, discussions about this work and some future researches are considered.

Section snippets

Overview of our approach

A topological map represents the robot environment by a graph. Paths are defined as sets of two distinct points, or “places” which must be detected and recognized by the robot using sensors data. These points provide the nodes of the map. Only a few relevant information about the places are required to locate and identify them. The edges between two nodes correspond to navigation operations such as wall following, visual servoing [8], …. These navigation operations take the robot from one node

Learning the topological graph

The exploration task consists in going over every path in the environment, memorizing the path connections in a GVG and learning some visual landmarks at the proximity to every meet point. From such a point at least two paths begin. Hereafter the different steps of the exploration task are listed.

Meet point detection. When it goes down an unexplored corridor, the robot is controlled to be on the GVG (between the two closer obstacles, typically the walls of a corridor); a meet point is detected

Learning landmarks

To learn a model of a landmark on an autonomous way, we need on the one hand to set criteria and methods to detect this landmark and on the other hand to build a model that will be reusable for recognition.

Experimental results

For the preliminary experiment in the Beckman Institute, as illustrated in Fig. 8, only sonars were used. Without vision, we have adopted a global localization technique based on odometer readings; we needed to minimize the number of wheel spins, and even with such a limitation, the node recognition procedure failed very often.

In Fig. 9, experimental results for a complex building environment are presented. The environment is not a regular corridor network; only a partial exploration has been

Discussion and future work

This paper has presented the integration of several topological based representations required for the navigation of a mobile robot in an office environment. It takes advantage both from the Generalized Voronoi Graph model, suitable to represent a network of corridors, and from a landmark-based topological map which has been proposed to get rid of the classical problems which occur with an explicit self-localization with respect to an absolute reference frame. We avoid the use of traditional

Prashanth Ranganathan was born in 1978 in India. He completed his high school at Sydney Technical High School, and then his Bachelors in Computer Engineering at University of Illinois at Urbana-Champaign. He did his undergraduate thesis in mobile robotics and navigation using sensory devices. He is currently with Microsoft TV as a software design engineer.

References (10)

  • H. Bulata, M. Devy, Incremental construction of a landmark-based and topological model of indoor environments by a...
  • H. Choset, K. Nagatani, A. Rizzi, Sensor based planning: using a honing strategy and local map method to implement the...
  • S. Koenig, R. Simmons, Unsupervised learning of probabilistic models for robot navigation, in: Proceedings of the IEEE...
  • D. Kortenkamp, T. Weymouth, Topological mapping for mobile robots using a combination of sonar and vision sensing, in:...
  • B. Kuipers, The spatial semantic hierarchy, Artificial Intelligence 119 (2000)...
There are more references available in the full text version of this article.

Cited by (0)

Prashanth Ranganathan was born in 1978 in India. He completed his high school at Sydney Technical High School, and then his Bachelors in Computer Engineering at University of Illinois at Urbana-Champaign. He did his undergraduate thesis in mobile robotics and navigation using sensory devices. He is currently with Microsoft TV as a software design engineer.

Jean-Bernard Hayet got his degree in Computer Science Engineering in 1999 from ENSTA (Paris, France). He also got a DEA degree in Artificial Intelligence and Pattern Recognition from Paris VI University and since 1999 he has been working at the LAAS-CNRS Robotics and Artificial Intelligence Group on a Ph.D. Thesis about the use of visual landmarks for mobile robots in indoor environments.

Michel Devy got his degree in Computer Science Engineering in 1976 from IMAG, in Grenoble (France). He received is Ph.D. in 1980 from LAAS-CNRS in Toulouse (France). Since 1980, he has participated to the Robotics and Artificial Intelligence Group of LAAS-CNRS; his research is devoted to the application of computer vision in Automation and Robotics. He has been involved in numerous national and international projects, about Manufacturing Applications, Mobile Robots for space exploration or for civil safety, 3D Vision for Intelligent Vehicles or Medical Applications. He is now Research Director at CNRS, head for the Perception Area in the Robotics Group of LAAS-CNRS and his main scientific topics concern Perception for Mobile Robots in natural or indoor environments.

Seth Hutchinson received his Ph.D. from Purdue University in 1988. He is an Associate Professor in the Department of Electrical and Computer Engineering at University of Illinois at Urbana-Champaign and a full-time Beckman Institute faculty member in the Artificial Intelligence Group. His fields of professional interest are robotics, computer vision, and artificial intelligence.

Frédéric Lerasle was born in France, in 1969. He received the Diplôme d’Ingénieur from the Centre Universitaires des Sciences et Techniques in 1992 and the Diplôme de Docteur-Ingénieur in 1997 from Blaise Pascal University, Clermont-Ferrand. He is currently a Maı̂tre de Conférences in Image Processing at the Paul Sabatier University in Toulouse. In 1997, he joined the Robotics and AI Group of the LAAS. His main fields of research include indoor robot navigation using perception and human–robot interaction.

View full text