A novel algorithm of extended neural networks for image recognition

https://doi.org/10.1016/j.engappai.2015.03.010Get rights and content

Abstract

As a class of important classifiers, feedforward neural networks (FNNs) have been used considerably in the study of pattern recognition. Since the inputs to FNNs are usually vectors, and many data are usually presented in the form of matrices, the matrices have to be decomposed into vectors before FNNs are employed. A drawback to this approach is that important information regarding correlations of elements within the original matrices are lost. Unlike traditional vector input based FNNs, a new algorithm of extended FNN with matrix inputs, called two-dimensional back-propagation (2D-BP), is proposed in this paper to classify matrix data directly, which utilizes the technique of incremental gradient descent to fully train the extended FNNs. These kinds of FNNs help to maintain the matrix structure of the 2D input features, which helps with image recognition. Promising experimental results of handwritten digits and face-image classification are provided to demonstrate the effectiveness of the proposed method.

Introduction

It is well-known that image recognition is a hot topic in the fields of machine learning and computer vision. In such recognition systems, many data images, such as handwritten digital images, face images, and palm images, are usually presented in the form of matrices. Clearly, determining how to classify these kinds of data is an important topic in pattern recognition.

Traditional image recognition system contains three steps: image pre-processing, feature extraction, and classification. At present, there are various methods for feature extraction, such as principal component analysis (PCA) (Jolliffe, 2002), independent component analysis (ICA) (Hyvärinen and Oja, 2000), and several popular applied classifiers, including K-nearest neighbourhoods (KNNs) (Shakhnarovich et al., 2008), support vector machines (SVMs) (Vapnik, 2000), feedforward neural networks (FNNs) (Hornik et al., 1989), and so on.

Nevertheless, these methods are usually based on vector inputs. Thus, when they are used in image processing, we first have to expand the matrix inputs into vector form. These types of transformation often lead to the loss of important information regarding the original matrix data, and thus affect the recognition process. On the other hand, expanding matrix inputs into vectors usually causes high dimensionality and increases the complexity of the used models.

To solve this problem, two-dimensional (2D) methods that directly operate on matrix data are proposed, for example, the two-dimensional principal component analysis (2DPCA) (Yang et al., 2004, Zhang and Zhou, 2005) and two-dimensional linear discriminant analysis (2DLDA) (Li and Yuan, 2005, Sanguansat et al., 2006, Yang et al., 2010), which were verified useful for extracting effective information about the inner structure of matrix data, as well as reducing the computational complexity of the extraction. It is natural to raise the question: For the existing vector-based classifiers, such as the SVM and FNN methods, can they be extended for matrix input?

Neural networks have played an important role in pattern recognition (Bishop, 1995, Lin et al., 1997, LeCun et al., 1998, Shang et al., 2006). In order to classify matrix data directly, and to preserve the matrix or 2D feature structure effectively, Lu et al. (2014) proposed a novel classifier, the two-dimensional neural network with random weights (2D-NNRW) method, and achieved good performance on face recognition. In fact, it is an extended 2D single hidden layer feed forward neural network (2D-SLFN) model that employs left and right projecting vectors to regress matrix inputs. Additionally, it uses the random idea to train the network, i.e. it randomly sets the left and right projecting vectors and the hidden biases, and then determines the output weights by solving a linear equation system. The results obtained in Lu et al. (2014) show that the use of 2D-SLFN improves face recognition accuracy.

The randomness in the NNRW algorithm can be understood in deeper detail when we consider the function approximation with Monte Carlo (MC) methods. It was shown in Igelnik and Pao (1995) that any continuous function defined on a compact set can be represented by a limit-integral of a multivariate continuous function that is integrated in parameter space. Although using the NNRW algorithm can simplify the learning steps taken by SLFNs, the following issues still remain in both the NNRW and 2D-NNRW methods:

  • The number of hidden nodes should be sufficiently large and supervised initialization is needed in order to model and compensate for the system׳s uncertainties.

  • There exists an over-fitting phenomenon caused by many additional hidden nodes in the NNRW method due to the MC approximating approach.

  • There is a predictive instability caused by randomly assignment of nonlinear weights, and the way of learning SLFNs with NNRW methods using singular value decomposition (SVD) usually produces large magnitude linear weights, which makes the networks highly sensitive to new data.

To overcome these issues, we have attempted to fully train the network with the aim of implementing incremental gradient based learning for 2D-SLFN. A learning algorithm called two-dimensional back-propagation (2D-BP) is proposed, where a momentum modification is added to improve convergence. A series of comparative studies of handwritten digits and face-image classification were carried out. The results of the testing datasets are promising, and support a positive statement regarding their performances among the 1D-BP and 2D-NNRW methods.

The rest of this paper is organized as follows. FNNs models and their corresponding training algorithms are reviewed in Section 2. A detailed description of our method is given in Section 3. An evaluation of the performance of our algorithm, the handwritten digits and face datasets employed in the experiments, and our results, which include comparisons and discussions, are presented in Section 4. Conclusions are presented in the final section.

Section snippets

Single hidden-layer feedforward neural networks

Generally, a SLFN is described as follows:f(x)=k=1Lβkg(wkx+bk)+α,where x is the input pattern vector, wk=[wk1,wk2,,wkd], bk are input layer weights and biases, respectively, and βk=[βk1,βk2,,βko] and α are the output layer weights and biases, respectively.

It is true that SLFNs are universal approximators (Hornik et al., 1989, Cybenko, 1989, Barron, 1993), even when the hidden-layer weights and bias are randomly assigned. In the case that weights and bias are randomly assigned, i.e., the

2D-BP for single hidden-layer feedforward neural networks

In order to maintain the matrix structure of the image inputs, 2D-SLFN is used in Lu et al. (2014), where the NNRW algorithm is extended to train this network, and the results suggest that 2D-SLFN is superior to general vector based SLFN (1D-SLFN) for image recognition. In this paper, we propose an incremental gradient based learning method for 2D-SLFN (2D-BP). Due to the left and right projecting vectors, the free weights of the 2D-SLFN are less than the general 1D-SLFN with the same number of

Performance evaluation

In this section we investigated the performance of the 2D-BP, and then compared it with two other methods: the 1D-BP and the 2D-NNRW.

Conclusions

In the structure of classical FNNs, the inputs are usually vectors, and thus FNNs cannot be used to classify matrix data directly. In this paper, a new FNNs algorithm, called the two-dimensional back-propagation (2D-BP), was proposed for classifying matrix data directly. The method of incremental gradient based learning is used to fully train the weights in extended 2D-FNNs. Since the proposed algorithm can retain important information of 2D input feature, it can be used as a classifier for

References (26)

  • Battiti, R., Masulli, F., 1990. BFGS optimization for faster and automated supervised learning. In: International...
  • C.M. Bishop

    Neural Networks for Pattern Recognition

    (1995)
  • G. Cybenko

    Approximation by superposition of sigmoidal function

    Math. Control, Signals Syst.

    (1989)
  • Cited by (12)

    • Delay independent robust stability analysis of delayed fractional quaternion-valued leaky integrator echo state neural networks with QUAD condition

      2019, Applied Mathematics and Computation
      Citation Excerpt :

      Real-valued Neural Networks (RVNN) have been successfully applied in practice, such as in modeling, control, associative memory and image recognition [1–6].

    • Super-resolution using neighbourhood regression with local structure prior

      2019, Signal Processing: Image Communication
      Citation Excerpt :

      It uses external LR and HR image pairs to estimate missing details. The main approaches here include sparse representation [20–24], neighbourhood regression [25–30], and deep learning [31–35], whereby impressive results have been obtained. Neighbourhood regression is a widely used method for SR and uses a certain number of nearest samples to explore the mapping relationship between LR and HR images.

    • Multistability analysis of delayed quaternion-valued neural networks with nonmonotonic piecewise nonlinear activation functions

      2019, Applied Mathematics and Computation
      Citation Excerpt :

      There is no doubt that the quaternion gradually shows its advantages in practical applications. Over the past few decades, the investigation on dynamic behaviors of neural networks has gained in popularity and plentiful results have been obtained for their useful applications in associative memory, pattern recognition, image processing, and so on (see [11–14] and references therein). As is well-known, the two-dimensional data can be well processed by complex-valued neural networks (CVNNs) or many real-valued neural networks (RVNNs).

    • Global μ-stability criteria for quaternion-valued neural networks with unbounded time-varying delays

      2016, Information Sciences
      Citation Excerpt :

      There is no deny that quaternion is an important topic needing much more attentions. Besides, neural networks have been one of the most interesting focuses due to the promising development and wide applications in diverse science and engineering fields such as associative memories [50], pattern recognition [8], control engineering [1]. So far, neural networks have been extensively investigated, see [4,44–46,51] and the references therein.

    View all citing articles on Scopus

    Supported by the National Natural Science Foundation of China (Nos. 61272023 and 91330118) and Zhejiang Provincial Natural Science Foundation of China (No. LY14A010027).

    View full text