Elsevier

Expert Systems with Applications

Volume 53, 1 July 2016, Pages 192-203
Expert Systems with Applications

Root-quatric mixture of experts for complex classification problems

https://doi.org/10.1016/j.eswa.2016.01.040Get rights and content

Highlights

  • We design a new ensemble system based on mixture of experts.

  • The anti-correlation measure is augmented to error function of mixture of experts.

  • The gating network assigns the weights to all of the output neurons of the experts.

  • The effect of anti-correlation measure is investigated.

  • Increasing anti-correlation measure will increase the diversity among the experts.

Abstract

Mixture of experts (ME) as an ensemble method consists of several experts and a gating network to decompose the input space into some subspaces regarding to the experts specialties. To increase the diversity between experts in ME, this paper incorporates a correlation penalty function into the error function of ME. The significant of this modification is providing an occasion to encourage experts to specialize on different parts of the input space and to create decorrelated experts. The experimental results of this approach reveals that the impacts of this penalty function is extremely improved the diversity of experts and the tradeoff between the accuracy and the diversity in ME. Moreover in the implementation of this method, the experts are trained simultaneously and they can communicate by the aid of the correlation penalty function. The performance of the proposed method on ten classification benchmark datasets shows that the average of accuracy of this method improves 1.94%, 3.7%, and 3.74% compared with the mixture of negatively correlated experts, ME and the negative correlation learning, respectively. Thus the proposed method can be considered as a better classifier for healthy and medical problems and also when the great non-stationary data should be classified.

Section snippets

Preliminaries and related works

Ensemble learning directs to combine multiple experts (classifiers or regressions) that are trained on a sample problem. Their decisions are combined to obtain better generalization ability in comparison with the base models. For this aim, ensemble learning applies diverse experts and tries to minimize their mistakes. Different methods are designed to create decorrelated (diverse) experts that can be classified as explicit and implicit methods (Brown & Yao, 2001). Implicit methods indirectly

Root-quartic mixture of experts (rtqrt-me) technique

In this section we introduce our proposed approach, root quartic mixture of experts (RTQRT-ME) that incorporates a penalty correlation measure into the error function of ME. The main aim of designing an ensemble system is improvement of the decision making phase by combining the outputs of different base experts. For this purpose, the individual experts are required to be decorrelated with one another. Incorporating the correlation penalty measure into the error function of system is one

Simulation results

In this section, the empirical results of the proposed method on benchmark datasets are presented to illustrate the performance of it.

Conclusion

In this paper, a novel ME based ensemble learning approach is presented. One of the most important problems in designing ensemble systems is the diversity between experts which affects the performance of the system. In order to create diversity among the experts, several methods have been developed, in some of which the experts are independently trained, with no interaction and cooperation. In the proposed method, RTQRT-ME, a penalty correlation function is incorporated into the error function

References (60)

  • LiuY. et al.

    Ensemble learning via negative correlation

    Neural Networks

    (1999)
  • LysiakR. et al.

    Optimal selection of ensemble classifiers using measures of competence and diversity of base classifiers

    Neurocomputing

    (2014)
  • MasoudniaS. et al.

    Combining features of negative correlation learning with mixture of experts in proposed ensemble methods

    Applied Soft Computing

    (2012)
  • MeoR. et al.

    LODE: A distance-based classifier built on ensembles of positive and negative observations

    Pattern Recognition

    (2012)
  • PeraltaB. et al.

    Embedded local feature selection within mixture of experts

    Information Sciences

    (2014)
  • RahmanA. et al.

    Ensemble classifier generation using non-uniform layered clustering and genetic algorithm

    Knowledge-Based Systems

    (2013)
  • SimidjievskiN. et al.

    Predicting long-term population dynamics with bagging and boosting of process-based models

    Expert Systems with Applications

    (2015)
  • UbeyliE.D.

    Wavelet/mixture of experts network structure for EEG signals classification

    Expert Systems with Applications

    (2008)
  • YoonJ.W. et al.

    Adaptive mixture-of-experts models for data glove interface with multiple users

    Expert Systems with Applications

    (2012)
  • ArmanoG. et al.

    Run-time performance analysis of the mixture of experts model

    Computer Recognition Systems

    (2011)
  • AroraR. et al.

    Comparative analysis of classification algorithms on different datasets using WEKA

    International Journal of Computer Applications

    (2012)
  • AsuncionA. et al.

    UCI machine learning

    Neural Computation

    (2007)
  • AvnimelechR. et al.

    Boosted mixture of experts: an ensemble learning scheme

    Neural Computation

    (1999)
  • BouchaffraD.

    Induced subgraph game for ensemble selection

  • BreimanL.

    Bagging predictors

    Machine Learning

    (1996)
  • BrownG. et al.

    On the effectiveness of negative correlation learning

  • CaoK.A.L. et al.

    Integrative mixture of experts to combine clinical factors and gene markers

    Bioinformatics

    (2010)
  • ChawlaN.V. et al.

    Smoteboost: improving prediction of the minority class in boosting

  • DemsarJ.

    Statistical comparisons of classifiers over multiple datasets

    Journal of Machine learning research

    (2006)
  • EbrahimpourR. et al.

    Boost-wise pre-loaded mixture of experts for classification tasks

    Neural Computing & Applications

    (2013)
  • Cited by (11)

    • Neural trees with peer-to-peer and server-to-client knowledge transferring models for high-dimensional data classification

      2019, Expert Systems with Applications
      Citation Excerpt :

      FELM-RT is a kind of mixture of experts that each expert is a neural tree exploiting ELM in the nodes. In the recent works on mixture of experts, one can see Abbasi, Shiri & Ghatee (2016a,b). In FELM-RT, the number of experts is equal to the number of clusters of features.

    • Classification using hierarchical mixture of discriminative learners: How to achieve high scores with few resources?

      2018, Expert Systems with Applications
      Citation Excerpt :

      In Kotsiantis (2011), they combine bagging, boosting, rotation forest and random subspace methods to classify the data. The most recent ensemble methods RTQRT-ME and R-RTQRT-ME (Abbasi et al., 2016b; 2016a) used MLP as weighting functions and experts and introduced a regularization term to increase the diversity of the experts. For this experiment, we used the same number of experts (five) as in Abbasi et al. (2016b, 2016a) and the structure of the tree is determined by a 5-cross validation.

    • A context aware system for driving style evaluation by an ensemble learning on smartphone sensors data

      2018, Transportation Research Part C: Emerging Technologies
      Citation Excerpt :

      It is possible to use reinforcement learning and unsupervised learning to adjust these thresholds automatically. Besides, some more modern learning algorithms such as mixture of experts (Abbasi et al., 2016) can be developed for solving this complex problem. To decrease battery consumption, we can use some feature selection methods or decreasing the frequency of sampling from sensors.

    • A regularized root-quartic mixture of experts for complex classification problems

      2016, Knowledge-Based Systems
      Citation Excerpt :

      In addition, as the mentioned in [24] to encourage the experts to learn different regions of input space, one can incorporate a correlation penalty function into the error function of the model, e.g., in [25], a negative correlation learning penalty term was incorporated into the error function of ME. In [26], a novel penalty correlation measure namely RTQRT-ME was incorporated into the error function to control the correlation between the experts. In addition, to improve the robustness of ME or RTQRT-ME in confronting with the overfitting and noise effects, in this paper, we add a regularization term to the error function of RTQRT-ME.

    • Semi-explicit mixture of experts based on information table

      2023, Journal of Ambient Intelligence and Humanized Computing
    View all citing articles on Scopus
    View full text