Elsevier

Neurocomputing

Volume 70, Issues 4–6, January 2007, Pages 896-903
Neurocomputing

Shape recognition based on neural networks trained by differential evolution algorithm

https://doi.org/10.1016/j.neucom.2006.10.026Get rights and content

Abstract

In this paper a new method for recognition of 2D occluded shapes based on neural networks using generalized differential evolution training algorithm is proposed. Firstly, a generalization strategy of differential evolution algorithm is introduced. And this global optimization algorithm is applied to train the multilayer perceptron neural networks. The proposed algorithms are evaluated through a plant species identification task involving 25 plant species. For this practical problem, a multiscale Fourier descriptors (MFDs) method is applied to the plant images to extract shape features. Finally, the experimental results show that our proposed GDE training method is feasible and efficient for large-scale shape recognition problem. Moreover, the experimental results illustrated that the GDE training algorithm combined with gradient-based training algorithms will achieve better convergence performance.

Introduction

The shape feature is one of the most important features for characterizing an object, which is commonly used in object recognition, matching and registration. In addition, the shape recognition is also an important part of machine intelligence that is useful for both decision-making and data processing. More importantly, the recognition-based on shape feature is also a central problem in those fields such as pattern recognition, image technology and computer vision, etc., which have received considerable attention recent years. Face recognition, image preprocessing, computer vision, fingerprint identification, handwriting analysis, and medical diagnosis, etc., are some of the common application areas of shape recognition. In particular, shape recognition has mutual effects with other research areas such as signal processing, neural networks, optimization theory, structural modeling and formal languages, etc. For shape recognition, there have been a wide range of methods proposed [1], [2]: structural methods organizing local features into graphs, trees, or strings; Fuzzy methods; Statistical methods; Transform methods, such as Fourier transform [3] or Hough transforms; Neural networks methods [4], [5], and so on. But most of approaches are confined to specific image types and require that all shapes must be preprocessed before recognition. However, an ever-increasing amount of image data in many application domains has generated additional requirements for real-time management and retrieval of images. Therefore, the emphasis on image recognition is not only on the accuracy, but also on the efficiency.

On the other hand, the neural network is widely used in pattern recognition kingdom as an effective classifier. Since the development of the back-propagation method, many algorithms have been proposed and used to train neural networks, such as modified back-propagation [6], back-propagation using the conjugate-gradient approach [7], scaled conjugate-gradient [8], the Levenberg–Marquadt algorithm [9]. The simulated annealing (SA) method and genetic algorithm (GA) method also have been proposed for network with non-differentiable transfer functions where the gradient information is not available [10], [11], [12]. Many of the existing training algorithms are suitable to small or middle scale networks structures and have a rather fast convergence speed. For the small-scale problem or small network structure, the gradient information usually is available and the training methods based on gradient information are rather fast and can convergence to global minima by repeated training and using randomly initialized weight values. However, for some large-scale real world problems, such as the above-mentioned shape recognition problem, many of them have worse performances than small or middle scale problems on convergence accuracy and speed. There are seldom suitable and reasonable network training algorithms when the neural network structures or the number of network parameters grow rapidly. For such large-scale neural networks structure, many of the training methods need an unacceptable computation cost in time and space. And the local minima problem also must be considered. The global optimization algorithms, such as GA and SA, may be useful to avoid such a local minima problem. In fact, there is no single training algorithm can have the best performance compared with all other methods on all problem domains. One feasible solution method is that use the global optimization algorithms combined with gradient information methods to train the neural network to achieve acceptable solution.

For the global optimization methods, GAs have been studied and found to be promising stochastic optimization methods. A survey and overview of GAs in evolving neural networks can be found in [13], [14]. Differential Evolution (DE) is one of the recent population-based global optimization techniques [15], [16]. Some works have applied DE to train neural networks [11], [12], in which the experimental data are small-scale problems. In this paper, a new generalization strategy of DE is applied to train feed-forward multi-layer perceptron neural networks (MLPNN) and compared with different type of training algorithms. Furthermore, this paper focuses on using the novel neural network-based method to perform shape recognition task through multiscale Fourier descriptors (MFDs) of shapes.

This paper is organized as follows: In Section 2, a generalization strategy of DE algorithm is described and discussed. In Section 3, a novel training method for neural network based on generalized DE algorithm is presented. In Section 4, the MFDs feature extraction method for shapes is presented. The experimental results are reported in Section 5, and Section 6 concludes the whole paper and gives related conclusions.

Section snippets

Generalization strategy of DE algorithm

DE is one of the recent population-based global optimization techniques [15], [16], which is a heuristic method for minimizing nonlinear and non-differentiable continuous space functions. The DE scheme entirely corresponds to a typical GA. But the principle difference consists in the mutation operation. In GA mutation is caused by small alterations of genes, whereas in DE Mutation is provided by combinations of individuals. The core of this operation is the formation of a difference vector,

GDE training algorithm

From the above, it is shown that DE is a heuristic method for minimizing nonlinear and non-differentiable continuous space functions, so it can be applied to global searches within the weight space of a typical neural network.

The most popular neural network model is the so-called MLPNN. Training an MLPNN to recognize objectors is typically realized by adopting an error correction strategy that adjusts the network weights through minimization of learning error:E=E(Y0,Y),where, Y is the real

MFDs of shape

In this section, the shape feature extraction method is described. Let Γ a planar curve defined byΓ={(x(u),y(u))|u[0,1]},where u is the normalized arc length parameter. The curvature κ of a planar curve is defined as the derivative of the tangent angle with respect to the arc length s. Then the curvature function of can be expressed as follows:κ(u)=x˙(u)y¨(u)-x¨(u)y˙(u)((x˙(u))2+(y˙(u))2)3/2,wherex˙(u)=dxdu,x¨(u)=d2xdu2,y˙(u)=dydu,y¨(u)=d2ydu2.

An evolved version of the curve is defined byΓσ={(X

Experimental results

In our work, a leaf image database is used in the following experiment, which was collected and built by ourselves in our lab. This database includes 25 species of different plants. Each species includes at least 40 leaves images, 20 of which are used as training samples. There are totally 2000 images with the database. A subset of the images (16 images from 4 different plant species) is shown in Fig. 2.

For each test image, 5 different scales were selected σ=20,40,60,80,100. And for each scale,

Conclusions

In this paper, a generalized differential evolution (GDE) training algorithm was used to train multilayer perceptron neural networks and the trained networks are used to shape recognition. Furthermore, a multiscale Fourier descriptors (MFDs) feature extraction method for shape recognition is proposed. For using the multiscale local features of shapes, it is much more robust and effective than other method using global features. The superiority of GDE training method to network training and MFD

Acknowledgement

This work was supported by the National Science Foundation of China (No. 60472111 and 60405002), the Postdoctoral Science Foundation of China (No. 20060390180) and the Scientific Research Foundation of Huaqiao University (No. 06BS217).

Ji-Xiang Du took courses as B.Sc. degree candidate in Vehicle Engineering, Hefei University of Technology, from September 1995 to July 1999, and obtained B.Sc. degree in July 1999. From September 1999 to July 2002, took courses as M.Sc. degree candidate in Vehicle Engineering, Hefei University of Technology, and obtained M.Sc. degree in July 2002. From February 2003 on, in pursuit for Ph.D. degree in Pattern Recognition and Intelligent System in University of Science and Technology of China

References (19)

  • S. Loncaric

    A survey of shape analysis techniques

    Pattern Recogn.

    (1998)
  • D. Zhang et al.

    Review of shape representation and description techniques

    Pattern Recogn.

    (2004)
  • G.P. Schmitz et al.

    Combinatorial evolution of regression nodes in feed forward neural networks

    Neural Networks

    (1999)
  • Y. Rui, A.C. She, T.S. Huang, Modified Fourier descriptors for shape representation-a practical approach, In: First...
  • G. Papadourakis et al.

    Curvature scale space driven object recognition with an indexing scheme based on artificial neural networks

    Pattern Recogn.

    (1999)
  • D.S. Huang

    Systematic Theory of Neural Networks for Pattern Recognition

    (1996)
  • L. Min, W. Shi-Xi, L. Yong-Hong, Fast learning algorithms for multi-layered feedforward neural network, In: Proceedings...
  • C. Charalambous

    Conjugate gradient algorithm for efficient training of artificial neural networks

    IEE G (Circuits, Dev. Syst.)

    (1992)
  • S. Doyle et al.

    Automated mirror design using an evolution strategy

    Opt. Eng.

    (1999)
There are more references available in the full text version of this article.

Cited by (140)

View all citing articles on Scopus

Ji-Xiang Du took courses as B.Sc. degree candidate in Vehicle Engineering, Hefei University of Technology, from September 1995 to July 1999, and obtained B.Sc. degree in July 1999. From September 1999 to July 2002, took courses as M.Sc. degree candidate in Vehicle Engineering, Hefei University of Technology, and obtained M.Sc. degree in July 2002. From February 2003 on, in pursuit for Ph.D. degree in Pattern Recognition and Intelligent System in University of Science and Technology of China (USTC), Hefei, China, and in December 2005, he received Ph.D. degree. Now, he is engaged in the postdoctoral study in Department of Automation of USTC.

De-Shuang Huang (SM’98) is a Professor and Ph.D. Advisor in the University of Science and Technology of China (USTC), Hefei, China, and Professor of Graduate School of the Institute of Intelligent Machines, Chinese Academy of Sciences (CAS). From September 2000 to March 2001, he worked as Research Associate in Hong Kong Polytechnic University. From April 2002 to June 2003, he worked as Research Fellow in City University of Hong Kong. From October to December 2003, he worked as Research Fellow in Hong Kong Polytechnic University. From July to December 2004, he worked as the University Fellow in Hong Kong Baptist University. Dr. Huang is a senior member of the IEEE.

Xiao-Feng Wang took courses as a bachelor's degree candidate in Computer Science Technology, Anhui University from September 1995 to July 1999, and obtained bachelor's degree in July 1999. From 1999 to 2002, he worked in Institute of Intelligent Machines, Chinese Academy of Sciences (CAS). From September 2002 to July 2005, took courses as a master's degree candidate in Pattern Recognition and Intelligent System, Institute of Intelligent Machines, and obtained master's degree in July 2005. Now, he is working in Department of Computer and Science Technology of Hefei University.

Xiao Gu obtained bachelor degree in Computer Science from Nanjing Normal University in July 2003. From September 2003 on, in pursuit for M.Sc. degree in Pattern Recognition and Intelligent System in Institute of Intelligent Machines, Chinese Academy of Sciences, Hefei. Her research interests: include Digital Image Processing, Software Exploitation, Neural Networks, etc.

View full text