Elsevier

Information Sciences

Volume 325, 20 December 2015, Pages 159-174
Information Sciences

Generalized type-2 fuzzy weight adjustment for backpropagation neural networks in time series prediction

https://doi.org/10.1016/j.ins.2015.07.020Get rights and content

Abstract

In this paper the comparison of a proposed neural network with generalized type-2 fuzzy weights (NNGT2FW) with respect to the monolithic neural network (NN) and the neural network with interval type-2 fuzzy weights (NNIT2FW) is presented. Generalized type-2 fuzzy inference systems are used to obtain the generalized type-2 fuzzy weights and are designed by a strategy of increasing and decreasing an epsilon variable for obtaining the different sizes of the footprint of uncertainty (FOU) for the generalized membership functions. The proposed method is based on recent approaches that handle weight adaptation using type-1 and type-2 fuzzy logic. The approach is applied to the prediction of the Mackey–Glass time series, and results are shown to outperform the results produced by other neural models. Gaussian noise was applied to the test data of the Mackey–Glass time series for finding out which of the presented methods in this paper shows better performance and tolerance to noise.

Introduction

One of the most important parts in the structure of the neural network can be considered to be the weights applied to the neurons, because these enable the learning process in the neural network. This part is the main focus of this paper, because the use of generalized type-2 fuzzy weights to obtain the values for the weights in the connections between the input and hidden layers, and the hidden and output layers for a neural network can improve the learning process. The proposed method is compared against a neural network with interval type-2 fuzzy weights using the same architecture and learning algorithms. Noise was also applied to the real data to analyze the performance of the methods under a higher level of uncertainty.

We propose the adaptation of the weights in the backpropagation algorithm for neural networks using generalized type-2 fuzzy inference systems [2]. This approach is different than the ones previously used in the literature, where the adaptation of the weights is made with a momentum variable and adaptive learning rate [4], [20], or using triangular or trapezoidal type-1 fuzzy numbers to describe the weights [28], [29]. Also, in previous work we presented interval type-2 fuzzy inference with triangular or Gaussians membership functions to obtain the weights for the neurons in the neural network [21], [22].

The proposed approach is applied to time series prediction for the Mackey–Glass time series. The objective of applying different forecasting models is obtaining the minimum prediction error for the data of the time series.

The paper is mainly based on comparing the performance for the neural network with generalized type-2 fuzzy weights against a neural network with interval type-2 fuzzy weights. This is an important issue to investigate; because the weights affect the learning process of the neural network and considering uncertainty in the weights of the neural network allow obtaining better results. We performed experiments with the two methods and presented the results and comparison of these methods for the prediction of the Mackey–Glass time series.

The main contribution is the proposed adaptation of the backpropagation algorithm to allow the neural network to manage data with uncertainty, using generalized type-2 fuzzy logic for obtaining generalized type-2 fuzzy weights in the connections between the neurons of the layers. This adaptation of the backpropagation algorithm using generalized type-2 fuzzy inference systems to generate the type-2 weights enables the neural network to handle data with noise, such as that of the Mackey–Glass time series and other complex time series.

The paper is structured as follows: the next section presents a background of different methods for managing the weights and modifications of the backpropagation algorithm in neural networks, and basic concepts of neural networks and generalized and interval type-2 fuzzy logic. Section 3 explains the proposed method and the problem description, the weights using generalized type-2 fuzzy systems, and the neural network architecture with generalized type-2 fuzzy weights proposed in this paper. Section 4 presents the simulation results for the proposed methods. Finally, in Section 5 the conclusions of this work are presented.

Section snippets

Neural networks

An artificial neural network is based on the processing of artificial neurons connections [14], [37]. The artificial neuron is composed of several elements as described below.

First we find the inputs and the weights for each input, these are connected to the neuron and then the neuron performs a summation of the multiplication of the input values for the weights (xiwij), and finally the activation function is used. This function is completed with the addition of a threshold amount i. This

Proposed method and problem description

The proposed approach presented in this paper is an enhancement of the traditional backpropagation algorithm in which we are now using generalized type-2 fuzzy sets to allow the neural network to handle data with uncertainty. In generalized type-2 fuzzy sets is necessary to systematically vary the footprint of uncertainty (FOU) of the membership functions to properly design the fuzzy systems for the corresponding applications.

The proposed method utilizes generalized type-2 fuzzy weights in the

Simulation results

The results obtained in the experiments using the neural network with generalized type-2 fuzzy weights are presented in Table 1 and Fig. 9. In this case all parameters of the neural network are established empirically. The best achieved prediction error is of 0.0548, and the average error is of 0.0714. We are presenting 10 experiments in Table 1, but the average error is calculated considering 40 experiments with the same parameters and conditions. The number of epochs for the network is of 100

Conclusions

In the experiments, when there is no noise present, we observe that the neural network with interval type-2 fuzzy weights obtains better results than the neural network with generalized type-2 fuzzy weights for the Mackey–Glass time series. However, the neural network with generalized type-2 fuzzy weights is better at different noise levels than the neural network with interval type-2 fuzzy weights and the monolithic neural network. This conclusion is based on the fact that for different levels

References (61)

  • HidalgoD. et al.

    An optimization method for designing type-2 fuzzy inference systems based on the footprint of uncertainty using genetic algorithms

    Expert Syst. Appl.

    (2012)
  • KamarthiS. et al.

    Accelerating neural network training using weight extrapolations

    Neural Netw.

    (1999)
  • KarnikN. et al.

    Applications of type-2 fuzzy logic systems to forecasting of time-series

    Inf. Sci.

    (1999)
  • LiuF.

    An efficient centroid type-reduction strategy for general type-2 fuzzy logic system

    Inf. Sci.

    (2008)
  • MeltserM. et al.

    Approximating functions by neural networks: a constructive solution in the uniform norm

    Neural Netw.

    (1996)
  • MoH. et al.

    Footprint of uncertainty for type-2 fuzzy sets

    Inf. Sci.

    (2014)
  • MollerM.F.

    A scaled conjugate gradient algorithm for fast supervised learning

    Neural Netw.

    (1993)
  • NevilleR.S. et al.

    Transformations of Sigma–Pi nets: obtaining reflected functions by reflecting weight matrices

    Neural Netw.

    (2002)
  • TungS.W. et al.

    eT2FIS: an evolving type-2 neural fuzzy inference system

    Inf. Sci.

    (2013)
  • YamJ. et al.

    Weight initialization method for improving training speed in feedforward neural network

    Neurocomputing

    (2000)
  • YeungD. et al.

    radial basis function network learning using localized generalization error bound

    Inf. Sci.

    (2009)
  • ZhaiD. et al.

    Uncertainty measures for general type-2 fuzzy sets

    Inf. Sci.

    (2011)
  • AbiyevR.

    A type-2 fuzzy wavelet neural network for time series prediction

    Lecture Notes in Computer Science

    (2010)
  • BealeE.M.L.

    A derivation of conjugate gradients

    Numerical Methods for Nonlinear Optimization

    (1972)
  • BelA. et al.

    Multiobjective optimization and comparison of nonsingleton type-1 and singleton interval type-2 fuzzy logic systems

    IEEE Trans. Fuzzy Syst.

    (2013)
  • CastilloO. et al.
    (2001)
  • CastilloO. et al.

    Type-2 Fuzzy Logic Theory and Applications

    (2008)
  • CastroJ. et al.

    An interval type-2 fuzzy logic toolbox for control applications

  • CastroJ. et al.

    An interval type-2 fuzzy neural network for chaotic time series prediction with cross-validation and akaike test

    Soft Computing for Intelligent Control and Mobile Robotics

    (2011)
  • CastroJ. et al.

    Building fuzzy inference systems with a new interval type-2 fuzzy logic toolbox

    Trans. Comput. Sci.

    (2008)
  • Cited by (55)

    View all citing articles on Scopus
    View full text