Elsevier

Information Fusion

Volume 2, Issue 3, September 2001, Pages 163-168
Information Fusion

Multiple classifiers combination by clustering and selection

https://doi.org/10.1016/S1566-2535(01)00033-1Get rights and content

Abstract

This paper proposes a novel algorithm for multiple classifiers combination based on clustering and selection technique (called M3CS), which can find in the feature space the regions where each classifier has best classification performance. The proposed method may be divided into two steps: clustering and selection (operation). At clustering step, the feature space is partitioned into several regions by clustering separately the correctly and incorrectly classified training samples from each classifier, and the performances of the classifier in each region are calculated. In the selection step, the most accurate classifier in the vicinity of the input sample is nominated to provide the final decision of the committee. The performance comparison between M3CS and Kuncheva's CS+DT method, as well as some simple aggregation methods such as maximum, minimum, average, and majority vote, confirms the validity of the proposed scheme.

Introduction

The aim of designing a pattern recognition system is to achieve the best possible classification performance for the given task. This objective traditionally leads to the development of different classifier designs, and the classifier with the best performance is selected as a final solution to the problem. However, it had been observed that different classifier designs, which potentially offered complementary information about the patterns to be classified [1], [5], [6], [7], [8], [9], could be used simultaneously to achieve considerably improved performance over that of the best individual classifier.

Algorithms for multiple classifiers combination may take two approaches: classifier fusion and classifier selection. In classifier fusion algorithms, all classifiers are supposed to be equally “experienced” in the whole feature space and their outputs are combined in some manner to achieve a consensus [9], [10], [11], [12]. Classifier selection scheme assumes that each classifier has expertise in some local regions of the feature space and attempts to determine which classifier is most likely to be correct for an unknown sample.

For the classifier selection scheme, a method of partitioning the feature space and estimating the performance of each classifier is required. In Woods' DCS-LA approach [2], the classification accuracy was estimated in small regions of feature space surrounding an unknown test sample, then the most locally accurate classifier was nominated to make the final decision. This method, however, was too time-consuming due to the accuracy estimation for each test sample. Kuncheva [3] presented an algorithm to statically select the best classifier. In this method, the training data were clustered to form the decision regions, and a confidence interval was used to determine whether one or multiple classifiers should be used to make a decision. However, the number of clusters must be predetermined, and the class labels of the training samples were disregarded in the clustering procedure.

In this paper, a clustering [4] and selection based algorithm (called M3CS) is proposed to integrate multiple classifiers. With each classifier, the training samples may be divided into correctly and incorrectly classified ones, which are then clustered, respectively, to form a partition of the feature space. Due to the difference between the classifiers' error characteristics, the partitions resulted from different classifiers are generally not the same. For a certain partition, the performance of the corresponding classifier in each region is estimated using training, or validation data. In the operation phase, the most accurate classifier in the vicinity of the input sample is appointed to make the final decision. See Fig. 1 in Section 2 for a geometrical view of the basic idea.

This paper is arranged as follows: First, the basic idea of the algorithm is introduced, then the iteration procedure is presented in detail. In Section 3, the data sets used for the experiment are described briefly. The experiment results are given in Section 4 and we conclude this paper in Section 5.

Section snippets

Feature space partition and classifier selection

The basic idea of the M3CS may be illustrated with an example, where two trained classifiers, denoted as F1 and F2, are available in the committee. According to these classifiers, two feature space partitions, as shown in Fig. 1, can be obtained. According to F1, the feature space is divided into six regions, say R1,R2,…,R6. The classification accuracy of F1 in these six regions is computed using the training or validation data. However, seven clusters are obtained from F2, and the performances

Data used

To evaluate the performance of M3CS, we select four data sets: Clouds, Phoneme, Satimage, and Waveform, from the ELENA project1 as well as the UCI data set repository [13]. These data sets are carefully selected so

Experiments and analysis

A series of experiments have been carried out to verify the performance of M3CS, and how it rates among some other aggregation methods such as average, maximum, minimum, majority vote and Kuncheva's CS+DT method. All these experiments are implemented on the four data sets introduced in the previous section, where the former 1500 samples in each data set are used for training while the remaining samples are used for testing. All the involved parameters, such as the parameters of the individual

Conclusions

Clustering and selection based multiple classifiers combination algorithm (M3CS) is proposed in this paper. First, the training data set is clustered according to each classifier's performance. Then, the output of the best classifier with response for the vicinity of the input sample is adopted as the final combination result. The performance of the proposed model is verified through the experiments. It should be noted that the fusion result is affected by the values of the parameters,

References (13)

There are more references available in the full text version of this article.

Cited by (51)

  • Multi-Manifold based Rotation Forest for classification

    2018, Applied Soft Computing Journal
    Citation Excerpt :

    Finally, Section 6 draws the conclusion. According to review of contributions in the field of ELSs, there are five methods for combining and generating the classifiers: CF [4,8], SCS [5,8,24], SES [7,8], DCS [10,11,25] and DES [8,9,24]. CF, SCS, and SES are static classifiers ensembles, while DCS and DES are dynamic.

  • Learning simultaneous adaptive clustering and classification via MOEA

    2016, Pattern Recognition
    Citation Excerpt :

    In this way, the classification learning benefits from the clustering learning, but it cannot overcome the difficulty of overtraining. Similar strategy is also adopted in [12]. In [13], a clustering-launched classification (CLC) is proposed.

  • Dynamic selection of the best base classifier in One versus One

    2015, Knowledge-Based Systems
    Citation Excerpt :

    Giacinto and Roli [19] also extend Woods’s work incorporating distance weighted and classifiers confidence levels to two new methods called A Priori and A Posteriori. On the other hand, there are also other works which are not based on the K-NN method, for instance, Liu and Yuan [28] propose to use clustering: they divide the feature space into several clusters for each base classifier. The unknown sample is assigned to a cluster for each base classifier, and the classifier of the most accurate cluster is selected to classify the unknown sample.

View all citing articles on Scopus

Supported by National Science Foundation Committee (No. 69789301).

View full text