Skip to main content
Log in

Reduced Multi-class Contour Preserving Classification

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

This research presents the augmentation of the original contour preserving classification technique to support multi-class data and to reduce the number of synthesized vectors, called multi-class outpost vectors (MCOVs). The technique has been proven to function on both synthetic-problem data sets and real-world data sets correctly. The technique also includes three methods to reduce the number of MCOVs by using minimum vector distance selection between fundamental multi-class outpost vectors and additional multi-class outpost vectors to select only MCOVs located at the decision boundary between consecutive classes of data. The three MCOV reduction methods include the FF-AA reduction method, the FA-AF reduction method, and the FAF-AFA reduction method. An evaluation has been conducted to show the reduction capability, the contour preservation capability, and the levels of classification accuracy of the three MCOV reduction methods on both non-overlapping and highly overlapping synthetic-problem data sets and highly overlapping real-world data sets. For non-overlapping problems, the experimental results present that the FA-AF reduction method can partially reduce the number of MCOVs while preserving the contour of the problem most accurately and obtaining similar levels of classification accuracy as when the whole set of MCOVs is used. For highly overlapping problems, the experimental results present that the FF-AA reduction method can partially reduce the number of MCOVs while preserving the contour of the problem most accurately and obtaining similar levels of classification accuracy as when the whole set of MCOVs is used.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19
Fig. 20
Fig. 21
Fig. 22
Fig. 23

Similar content being viewed by others

References

  1. Haykin S (1999) Neural networks: a comprehensive foundation, 2nd edn. Prentice Hall, Upper Saddle River

    MATH  Google Scholar 

  2. Russell S, Norving P (2004) Artificial intelligence: a modern approach, 2nd edn. Pearson Education, Delhi

    Google Scholar 

  3. Negnevitsky M (2005) Artificial intelligence: a guide to intelligent systems, 2nd edn. Addison Wesley, Essex

    Google Scholar 

  4. Tanprasert T, Tanprasert C, Lursinsap C (1998) Contour preserving classification for maximal reliability. In: International joint conference on neural networks, pp 1125–1130. doi:10.1109/IJCNN.1998.685930

  5. Chun RK, McNamee LP (1990) Immunization on neural networks against hardware faults. In: International symposium on circuits and systems, pp 714–718, doi:10.1109/ISCAS.1990.112179

  6. Lursinsap C, Tanprasert T (1997) Fault immunization technique for artificial neural network. In: International conference on neural networks, pp 302–307. doi:10.1109/ICNN.1997.611683

  7. Clay RD, Sequin CH (1992) Fault tolerance training improves generalization and robustness. In: International joint conference on neural networks, pp 769–774. doi:10.1109/IJCNN.1992.287094

  8. Minnix JI (1992) Fault tolerance of backpropagation neural network trained on noisy inputs. In: International joint conference on neural networks, pp 847–852. doi:10.1109/IJCNN.1992.287081

  9. Chakraborty G, Noguchi S (1996) Improving generalization of a well trained network. In: International conference on neural networks, pp 276–281. doi:10.1109/ICNN.1996.548904

  10. Fuangkhon P, Tanprasert T (2009) An incremental learning algorithm for supervised neural network with contour preserving classification. In: International conference on electrical engineering/electronics, computer, telecommunications and information technology, pp 470–473. doi:10.1109/ECTICON.2009.5137153

  11. Fuangkhon P, Tanprasert T (2009) An adaptive learning algorithm for supervised neural network with contour preserving classification. In: International conference on artificial intelligence and computational intelligence, pp 389–398. doi:10.1007/978-3-642-05253-8-43

  12. Fuangkhon P, Tanprasert T (2010) A dual network adaptive learning algorithm for supervised neural network with contour preserving classification for soft real time applications. In: International conference on hybrid artificial intelligence system, pp 128–135. doi:10.1007/978-3-642-13769-3-16

  13. Fuangkhon P (2014) An incremental learning preprocessor for feed-forward neural network. Artific Intell Rev 41(2):183–210. doi:10.1007/s10462-011-9304-0

    Article  Google Scholar 

  14. Mongkonsirivatana J (2005) Neural network reliability enhancement with boundary detection contour preserving training. In: International conference on intelligent technology, pp 234–239

  15. Fuangkhon P, Tanprasert T (2012) Multi-class contour preserving classification. In: International conference on intelligent data engineering and automated learning, pp 35–42. doi:10.1007/978-3-642-32639-4-5

  16. Breiman L (1998) Half and half bagging and hard boundary points, Technical Report, Statistics Department, University of California, Berkeley

  17. Hart P (1968) The condensed nearest neighbor rule. IEEE Trans Inf Theory 14(3):515–516. doi:10.1109/TIT.1968.1054155

    Article  Google Scholar 

  18. Ritter GL, Woodruff HB, Lowry SR, Isenhour T (1975) An algorithm for a selective nearest neighbor decision rule. IEEE Trans Inf Theory 21(6):665–669. doi:10.1109/TIT.1975.1055464

    Article  MATH  Google Scholar 

  19. Gates G (1972) The reduced nearest neighbor rule. IEEE Trans Inf Theory 18(3):431–433. doi:10.1109/TIT.1972.1054809

    Article  Google Scholar 

  20. Dasarathy BV, Sánchez JS, Townsend S (2000) Nearest neighbor editing and condensing tools-synergy exploitation. Pattern Anal Appl 3(1):19–30

    Article  Google Scholar 

  21. Wilson D (1972) Asymptotic properties of nearest neighbor rules using edited data. IEEE Trans Syst Man Cybern 2(3):408–421. doi:10.1109/TSMC.1972.4309137

    Article  MathSciNet  MATH  Google Scholar 

  22. Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn 20(3):273–297. doi:10.1023/A:1022627411411

    MATH  Google Scholar 

  23. Verleysen M, Bodt E, Wertz V (2014) Clouds database, UCL enhanced learning for evolutive neural architectures. https://www.elen.ucl.ac.be/neural-nets/Research/Projects/ELENA/elena.htm

  24. Verleysen M, Bodt E, Wertz V (2014) Gaussian 2D–8D databases, UCL enhanced learning for evolutive neural architectures. https://www.elen.ucl.ac.be/neural-nets/Research/Projects/ELENA/elena.htm

  25. Kohavi R, Becker B (2012) Adult income, UCI machine learning repository. http://archive.ics.uci.edu/ml/datasets/Adult

  26. Srinivasan A (2012) Statlog (Landsat Satellite), UCI machine learning repository. http://archive.ics.uci.edu/ml/datasets/Statlog+(Landsat+Satellite)

  27. Cestnik B (2012) Statlog (Shuttle), UCI machine learning repository. http://archive.ics.uci.edu/ml/datasets/Statlog+(Shuttle)

  28. Blackard JA, Dean DJ, Anderson CW (2012) Forest cover type, UCI machine learning repository. http://archive.ics.uci.edu/ml/datasets/Covertype

  29. Alpaydin E, Alimoglu F (2012) Pen-based recognition of handwritten digits, UCI machine learning repository. http://archive.ics.uci.edu/ml/datasets/Pen-Based+Recognition+of+Handwritten+Digits

  30. Alpaydin E, Kaynak C (2012) Optical recognition of handwritten digits, UCI machine learning repository. http://archive.ics.uci.edu/ml/datasets/Optical+Recognition+of+Handwritten+Digits

  31. Verleysen M, Bodt E, Wertz V (2014) Iris database, UCL enhanced learning for evolutive neural architectures. https://www.elen.ucl.ac.be/neural-nets/Research/Projects/ELENA/elena.htm

  32. Verleysen M, Bodt E, Wertz V (2014) Phoneme database, UCL enhanced learning for evolutive neural architectures, https://www.elen.ucl.ac.be/neural-nets/Research/Projects/ELENA/elena.htm

  33. Verleysen M, Bodt E, Wertz V (2014) Satimage database, UCL enhanced learning for evolutive neural architectures. https://www.elen.ucl.ac.be/neural-nets/Research/Projects/ELENA/elena.htm

  34. Verleysen M, Bodt E, Wertz V (2014) Texture database, UCL enhanced learning for evolutive neural architectures. https://www.elen.ucl.ac.be/neural-nets/Research/Projects/ELENA/elena.htm

  35. Chang CC, Lin CJ (2011) LIBSVM: a library for support vector machines. ACM Trans Intell Syst Technol 2(27):1–27. doi:10.1145/1961189.1961199

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Piyabute Fuangkhon.

Appendix

Appendix

1.1 Runtime Complexity: Multi-class Contour Preserving Classification

The runtime complexity of MCOV Generation Algorithm (\(T_{mcov(n)}\)) comprises the runtime complexity of FMCOV Generation Algorithm (\(T_{fmcov(n)}\)) and the runtime complexity of AMCOV Generation Algorithm (\(T_{amcov(n)}\)) as described in Eq. (19).

$$\begin{aligned} T_{mcov(n)} = T_{fmcov(n)} + T_{amcov(n)}. \end{aligned}$$
(19)

where n is the number of input vectors in all classes.

The FMCOVs and AMCOVs are generated from the algorithm presented in Algorithm 1 using the Euclidean distance function described in Eq. (24) on a set of input vectors. The runtime complexity of Euclidean distance is described in Eq. (25).

The runtime complexity of the FMCOV Generation Algorithm (\(T_{fmcov(n)}\)) is described in Eq. (20).

$$\begin{aligned} T_{fmcov(n)}= & {} O\left( \sum _{z=1}^{c} (n_{z} \times \left( \sum _{z=1}^{c} (n_{z}) - n_{z})\right) \times T_{d(p,q)}\right) .\nonumber \\= & {} O\left( \sum _{z=1}^{c} \left( n_{z} \times (cn - n_{z})\right) \times d\right) .\nonumber \\= & {} O\left( c(c-1)n^2 \times d\right) .\nonumber \\= & {} O\left( c^2n^2\right) .\nonumber \\= & {} O\left( n^2\right) . \end{aligned}$$
(20)

where \(n_{z}\) is the number of input vectors in class z, n is the number of input vectors in all classes, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector, k is a constant and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

The runtime complexity of the AMCOV Generation Algorithm (\(T_{amcov(n)}\)) is described in Eq. (21).

$$\begin{aligned} T_{amcov(n)}= & {} O\left( \sum _{z=1}^{c} n_{z} \times T_{d(p,q)}\right) .\nonumber \\= & {} O(cn \times d).\nonumber \\= & {} O(n). \end{aligned}$$
(21)

where \(n_{z}\) is the number of input vectors in class z, n is the number of input vectors in all classes, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

As a result, the runtime complexity of the MCOV Generation Algorithm (\(T_{mcov(n)}\)) described in Eq. (19) is formulated as Eq. (22).

$$\begin{aligned} T_{mcov(n)}= & {} T_{fmcov(n)} + T_{amcov(n)}.\nonumber \\= & {} O(n^2 + n).\nonumber \\= & {} O(n^2). \end{aligned}$$
(22)

where n is the number of input vectors in all classes.

1.2 Runtime Complexity: Reduced Multi-class Contour Preserving Classification

The runtime complexity of the RMCOV Generation Algorithm (\(T_{rmcov(m+n)}\)) comprises the runtime complexity of the RFMCOV Generation Algorithm (\(T_{rfmcov(m)}\)) and the runtime complexity of RAMCOV Generation Algorithm (\(T_{ramcov(n)}\)) as described in Eq. (23).

$$\begin{aligned} T_{\textit{rmcov}(m+n)} = T_{\textit{rfmcov}(m)} + T_{\textit{ramcov}(n)}. \end{aligned}$$
(23)

where m is the number of FMCOVs in all classes and n is the number of AMCOVs in all classes.

Since there are three MCOV reduction methods for selecting FMCOVs and AMCOVs as shown in Sect. 3.3, the runtime complexity of RFMCOV Generation Algorithm and the runtime complexity of RAMCOV Generation Algorithm of each reduction method is different from any other reduction methods. “FF-AA Reduction Method” section explains the runtime complexity of the FF-AA reduction method. “FA-AF Reduction Method” section explains the runtime complexity of the FA-AF reduction method. “FAF-AFA Reduction Method” section explains the runtime complexity of the FAF-AFA reduction method.

1.2.1 FF-AA Reduction Method

This section presents the Runtime Complexity of the FF-AA Reduced Multi-class Outpost Vector Generation Algorithm: FOV vs FOV, AOV vs AOV.

The FF-AA RFMCOVs and FF-AA RAMCOVs are selected from the algorithm presented in Algorithm 3 using the Euclidean distance function described in Eq. (24) on a set of FMCOVs and a set of AMCOVs. The runtime complexity of Euclidean distance is described in Eq. (25).

The runtime complexity of the FF-AA RFMCOV Generation Algorithm (\(T_{rfmcov(m)}\)) is described in Eq. (26).

$$\begin{aligned} d(p,q)= & {} \sqrt{\left( p_{1}-q_{1}\right) ^2 + \left( p_{2}-q_{2}\right) ^2 + \cdots + \left( p_{f}-q_{f}\right) ^2}.\nonumber \\= & {} \sqrt{\sum _{i=1}^{f} \left( p_{i}-q_{i}\right) ^2}. \end{aligned}$$
(24)

where p and q are two points in Euclidean space, \(p_i\) and \(q_i\) are the value of the feature i of input vectors p and q, respectively and f is the number of features of the point.

$$\begin{aligned} T_{d(p,q)}= & {} O\left( \sqrt{\sum _{i=1}^{d} \left( p_{i}-q_{i}\right) ^2}\right) .\nonumber \\= & {} O(d). \end{aligned}$$
(25)

where p and q are two points in Euclidean space, \(p_i\) and \(q_i\) are the value of the dimension i of input vectors p and q, respectively and d is the number of dimensions of the point.

$$\begin{aligned} T_{rfmcov(m)}= & {} O\left( \sum _{z=1}^{c} (m_{z} \times \left( \sum _{z=1}^{c} (m_{z}) - m_{z})\right) \times T_{d(p,q)}\right) .\nonumber \\= & {} O\left( \sum _{z=1}^{c} \left( m_{z} \times (cm - m_{z})\right) \times d\right) .\nonumber \\= & {} O\left( c(c-1)m^2 \times d\right) .\nonumber \\= & {} O\left( m^2\right) . \end{aligned}$$
(26)

where \(m_{z}\) is the number of FMCOVs in class z, m is the number of all FMCOVs, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector, k is a constant and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

The runtime complexity of the FF-AA RAMCOV Generation Algorithm (\(T_{ramcov(n)}\)) is described in Eq. (27).

$$\begin{aligned} T_{ramcov(n)}= & {} O\left( \sum _{z=1}^{c} (n_{z} \times \left( \sum _{z=1}^{c} (n_{z}) - n_{z})\right) \times T_{d(p,q)}\right) .\nonumber \\= & {} O\left( \sum _{z=1}^{c} \left( n_{z} \times (cn - n_{z})\right) \times d\right) .\nonumber \\= & {} O\left( c(c-1)n^2 \times d\right) .\nonumber \\= & {} O\left( n^2\right) . \end{aligned}$$
(27)

where \(n_{z}\) is the number of AMCOVs in class z, n is the number of all AMCOVs, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector, k is a constant and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

As a result, the runtime complexity of the FF-AA Reduced Multi-class Outpost Vector Generation Algorithm (\(T_{rmcov(n)}\)) described in Eq. (23) is formulated as Eq. (28).

$$\begin{aligned} T_{rmcov(m+n)}= & {} T_{rfmcov(m)} + T_{ramcov(n)}.\nonumber \\= & {} O\left( m^2 + n^2\right) . \end{aligned}$$
(28)

where m is the number of all FMCOVs and n is the number of all AMCOVs.

Since the number of FMCOVs and the number of AMCOVs are generally equal to the number of input vectors, m is generally equal to n as well. When m is equal to n, the runtime complexity of FF-AA reduction method will be equal to \(O(m^2)\).

1.2.2 FA-AF Reduction Method

This section presents the Runtime Complexity of the FA-AF RMCOV Generation Algorithm: FOV vs AOV, AOV vs FOV.

The FA-AF RFMCOVs and FA-AF RAMCOVs are selected from the algorithm presented in Algorithm 4 using the Euclidean distance function described in Eq. (24) on a set of FMCOVs and a set of AMCOVs. The runtime complexity of Euclidean distance is described in Eq. (25).

The runtime complexity of the FA-AF RFMCOV Generation Algorithm (\(T_{rfmcov(m)}\)) is described in Eq. (29).

$$\begin{aligned} T_{rfmcov(m)}= & {} O\left( \sum _{z=1}^{c} (m_{z} \times \left( \sum _{z=1}^{c} (n_{z}) - n_{z})\right) \times T_{d(p,q)}\right) .\nonumber \\= & {} O\left( \sum _{z=1}^{c} (m_{z} \times (cn - n_{z})) \times d\right) .\nonumber \\= & {} O(c(c-1)mn \times d).\nonumber \\= & {} O(mn). \end{aligned}$$
(29)

where \(m_{z}\) is the number of FMCOVs in class z, \(n_{z}\) is the number of AMCOVs in class z, m is the number of all FMCOVs, n is the number of all AMCOVs, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector, k is a constant and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

The runtime complexity of the FA-AF RAMCOV Generation Algorithm (\(T_{ramcov(n)}\)) is described in Eq. (30).

$$\begin{aligned} T_{ramcov(n)}= & {} O\left( \sum _{z=1}^{c} (n_{z} \times \left( \sum _{z=1}^{c} (m_{z}) - m_{z})\right) \times T_{d(p,q)}\right) .\nonumber \\= & {} O\left( \sum _{z=1}^{c} (n_{z} \times (cm - m_{z})) \times d\right) .\nonumber \\= & {} O(c(c-1)nm \times d).\nonumber \\= & {} O(nm).\nonumber \\ \end{aligned}$$
(30)

where \(m_{z}\) is the number of FMCOVs in class z, \(n_{z}\) is the number of AMCOVs in class z, m is the number of all FMCOVs, n is the number of all AMCOVs, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector, k is a constant and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

As a result, the runtime complexity of the FA-AF RMCOV Generation Algorithm (\(T_{rmcov(n)}\)) described in Eq. (23) is formulated as Eq. (31).

$$\begin{aligned} T_{rmcov(m+n)}= & {} T_{rfmcov(m)} + T_{ramcov(n)}.\nonumber \\= & {} O(mn + nm).\nonumber \\= & {} O(mn). \end{aligned}$$
(31)

where m is the number of all FMCOVs and n is the number of all AMCOVs.

Since the number of FMCOVs and the number of AMCOVs are generally equal to the number of input vectors, m is generally equal to n as well. When m is equal to n, the runtime complexity of FA-AF reduction method will be equal to \(O(m^2)\).

1.2.3 FAF-AFA Reduction Method

This section presents the Runtime Complexity of the FAF-AFA RMCOV Generation Algorithm: FOV vs FOV+AOV, AOV vs AOV+FOV.

The FAF-AFA RFMCOVs and FAF-AFA RAMCOVs are selected from the algorithm presented in Algorithm 5 using the Euclidean distance function described in Eq. (24) on a set of FMCOVs and a set of AMCOVs. The runtime complexity of Euclidean distance is described in Eq. (25).

The runtime complexity of the FAF-AFA RFMCOV Generation Algorithm (\(T_{rfmcov(m)}\)) is described in Eq. (32).

$$\begin{aligned} T_{rfmcov(m)}= & {} O\left( \sum _{z=1}^{c} \left( m_{z} \times \left( (\sum _{z=1}^{c} (n_{z}) - n_{z}) + \left( \sum _{z=1}^{c} (m_{z}) - m_{z}\right) \right) \right) \times T_{d(p,q)}\right) .\nonumber \\= & {} O\left( \sum _{z=1}^{c} \left( m_{z} \times \left( \left( cn - n_{z}\right) + \left( cm - m_{z}\right) \right) \times d\right) \right) .\nonumber \\= & {} O\left( (c(c-1)mn) + (c(c-1)m^2) \times d\right) .\nonumber \\= & {} O\left( \left( c^2mn + c^2m^2\right) \times d\right) .\nonumber \\= & {} O\left( mn + m^2\right) . \end{aligned}$$
(32)

where \(m_{z}\) is the number of FMCOVs in class z, \(n_{z}\) is the number of AMCOVs in class z, m is the number of all FMCOVs, n is the number of all AMCOVs, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector, k is a constant and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

The runtime complexity of the FAF-AFA RAMCOV Generation Algorithm (\(T_{ramcov(n)}\)) is described in Eq. (33).

$$\begin{aligned} T_{ramcov(n)}= & {} O\left( \sum _{z=1}^{c} \left( n_{z} \times \left( \left( \sum _{z=1}^{c} (n_{z}) - n_{z}\right) + \left( \sum _{z=1}^{c} (m_{z}) - m_{z}\right) \right) \right) \times T_{d(p,q)}\right) .\nonumber \\= & {} O\left( \sum _{z=1}^{c} (n_{z} \times ((cn - n_{z}) + (cm - m_{z})) \times d\right) .\nonumber \\= & {} O\left( \left( c(c-1)n^2\right) + (c(c-1)nm) \times d\right) .\nonumber \\= & {} O\left( \left( c^2n^2 + c^2nm\right) \times d\right) .\nonumber \\= & {} O\left( n^2 + nm\right) . \end{aligned}$$
(33)

where \(m_{z}\) is the number of FMCOVs in class z, \(n_{z}\) is the number of AMCOVs in class z, m is the number of all FMCOVs, n is the number of all AMCOVs, c is the number of classes of data, p and q are two points in different classes, d is the number of dimensions of the input vector, k is a constant and \(T_{d(p,q)}\) is the runtime complexity of Euclidean distance function between points p and q.

As a result, the runtime complexity of the FAF-AFA RMCOV Generation Algorithm (\(T_{rmcov(n)}\)) described in Eq. (23) is formulated as Eq. (34).

$$\begin{aligned} T_{rmcov(m+n)}= & {} T_{rfmcov(m)} + T_{ramcov(n)}.\nonumber \\= & {} O\left( \left( mn + m^2\right) + \left( n^2 + nm\right) \right) .\nonumber \\= & {} O\left( m^2 + n^2\right) . \end{aligned}$$
(34)

where m is the number of all FMCOVs and n is the number of all AMCOVs.

Since the number of FMCOVs and the number of AMCOVs are generally equal to the number of input vectors, m is generally equal to n as well. When m is equal to n, the runtime complexity of FAF-AFA reduction method will be equal to \(O(m^2)\). All three MCOV reduction methods have identical runtime complexity.

1.3 Experiment: Support Vector Machine (SVM)

The experiments in Sects. 4.14.2, and 4.3 are conducted on a feed-forward neural network to show the practicality of the three MCOV reduction methods in terms of MCOV reduction capability and generalization on a non-overlapping four-class synthetic-problem data set, eight highly overlapping two-class synthetic-problem data sets from the ELENA Project, and ten real-world data sets from both UCI machine learning repository and ELENA Project, respectively. This section additionally presents a comparison on the generalization among the three MCOV reduction methods and the Support Vector Machine (SVM) [22] on these data sets. For references, the characteristics of a non-overlapping four-class synthetic-problem data set are described in the first paragraph of Sect. 4.1, the characteristics of eight highly overlapping two-class synthetic-problem data sets from the ELENA Project are presented in Table 4, the characteristics of ten real-world data sets from both UCI machine learning repository and ELENA Project are presented in Table 8. The experiments are conducted with a LIBSVM [35] on MATLAB R2014b using the following parameters:

  • SVM Type = C-SVC (multi-class classification)

  • Kernel Function = Radial Basis Function

  • Degree in Kernel Function = 3

  • Gamma in Kernel Function = 1

  • Coefficient 0 in Kernel Function = 0

  • Cost of C-SVC = 1

  • Weight of C-SVC = 1

  • Cache Memory Size = 100 MB

  • Tolerance of Termination Criterion = 0.001

  • Shrinking Heuristics = 1

  • Probability Estimates = 0

Table 12 Misclassification rates of three MCOV reduction methods and Support Vector Machine on both synthetic and real-world data sets

Table 12 presents the misclassification rates of the three MCOV reduction methods and SVM on both synthetic and real-world data sets. The experimental results from a feed-forward neural network on ORG+FF-AA, ORG+FA-AF, and ORG+FAF-AFA data sets are taken from Tables 25610, and 11, respectively. The experiment result from the method that yields the lowest misclassification rate is highlighted in bold. For non-overlapping synthetic-problem data set having various population sizes, the three MCOV reduction methods yield significantly lower misclassification rates than the SVM. For overlapping synthetic-problem data sets, the FF-AA reduction method yields slightly lower misclassification rates than the SVM, and much lower misclassification rates than FA-AF and FAF-AFA reduction methods. More precisely, it yields the lowest misclassification rates. For overlapping real-world data sets, the three MCOV reduction methods generally yield lower misclassification rates than the SVM. It is noticeable that the misclassification rates of SVM on Statlog (Landsat Satellite), Forest Cover Type, Pen-based Recognition of Handwritten Digits, Optical Recognition of Handwritten Digits, and Satimage data sets are significantly large. SVM might need to be fine-tuned to be able to classify these data sets more accurately, or it may not be suitable for these data sets.

Based on the results from these experiments, it can be concluded that the three MCOV reduction methods outperform a Support Vector Machine on both synthetic and real-world data sets.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Fuangkhon, P., Tanprasert, T. Reduced Multi-class Contour Preserving Classification. Neural Process Lett 43, 759–804 (2016). https://doi.org/10.1007/s11063-015-9446-1

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-015-9446-1

Keywords

Navigation