Reducing samples for accelerating multikernel semiparametric support vector regression

https://doi.org/10.1016/j.eswa.2009.12.058Get rights and content

Abstract

In this paper, the reducing samples strategy instead of classical ν-support vector regression (ν-SVR), viz. single kernel ν-SVR, is utilized to select training samples for admissible functions so as to curtail the computational complexity. The proposed multikernel learning algorithm, namely reducing samples based multikernel semiparametric support vector regression (RS-MSSVR), has an advantage over the single kernel support vector regression (classical ε-SVR) in regression accuracy. Meantime, in comparison with multikernel semiparametric support vector regression (MSSVR), the algorithm is also favorable for computational complexity with the comparable generalization performance. Finally, the efficacy and feasibility of RS-MSSVR are corroborated by experiments on the synthetic and real-world benchmark data sets.

Section snippets

Motivation

The support vector machines (SVMs) proposed by Vapnik and his group (Burges, 1998, Cristianini and Shawe-Taylor, 2000, Schölkopf and Smola, 2002, Vapnik, 1995) have a foolproof theoretical foundation, viz. structural risk minimization (SRM) principle, which minimizes the upper bound of generalization error consisting of training errors and the confidence interval. As a state-of-the-art tool, SVM was first presented to cope with binary classification problem, and then it was extended to

Multikernel semiparametric ε-SVR

Considering the training set {(xi,di)}i=1N, where xiRn is the input variable and diR is the corresponding output variable, with the ε-insensitive loss function, we can get the following modelminw,b12wsTws+Ci=1N(ξi+ξi)s.t.di-ws·φs(xi)-p=1Bbpϕp(xi)ε+ξiws·φs(xi)+p=1Bbpϕp(xi)-diε+ξiξi,ξi0,i=1,,Nwhere ε>0 is the width of the tolerance band, C>0 is the user-selected regularization parameter, ws represents the model complexity, φs(·) is usually a nonlinear mapping which is induced from the

Reducing samples based MSSVR

Based on experimental observations that support vectors tend to take the extreme target values among the values of their k-nearest neighbors, Guo and Zhang (2007) proposed a reducing samples strategy to alleviate the training computational burden of SVR with the comparable generalization performance with that trained on the full training set. Under some conditions, they gave 4 propositions mathematically to support their viewpoint. In addition, plenty of synthetic and real-world experiments

Experiments

To validate the effectiveness and feasibility of Algorithm 3, in this section, we will do experiments on the synthetic and real-world benchmark data sets. In our paper, all the experiments are performed on a personal computer with AMD 3200+ (2.01GHz) processor, 512MB memory, and Windows XP operation system in a MATLAB 7.1 environment. For all algorithms, the quadratic programming is solved using the active method in the MATLAB toolbox. Moreover, to conveniently compare with different

Conclusions

In our real world, a lot of systems own different data trends in different regions. In this situation, the commonly-used single-kernel learning algorithm sometimes does not achieve satisfactory result. Hence, it is necessary to develop multikernel learning algorithms. Nguyen and Tay (2008) proposed a multikernel semiparametric support vector regression to cope with the systems holding complicated structure, viz. different data trends in different regions. Compared with the single-kernel

Acknowledgment

This research was supported by the National Natural Science Foundation of China under Grant No. 50576033.

References (27)

  • N. Cristianini et al.

    An introduction to support vector machines

    (2000)
  • Joachims, T. (1999). Making large-scale SVM learning practical. In Advances in kernel methods-support vector machine....
  • D. Kim et al.

    ε-tube based pattern selection for support vector machines

    Lecture Notes in Computer Science

    (2006)
  • Cited by (8)

    • Selecting rows and columns for training support vector regression models with large retail datasets

      2013, European Journal of Operational Research
      Citation Excerpt :

      However, business data sets with hundreds of input variables are very sparse due to curse of dimensionality and the distance definition affects the choice of training data. Zhao et al. (2010) choose observations within the multiple kernel support vector regression framework by extending Guo and Zhang’s ideas: they use the Euclidean distance between two observations or the cosine of the angle between the vectors, while noting that the measurement function can be defined in different ways. Kim and Cho (2006) create bootstrap samples to estimate the regression function, and retain those observations with high probability of falling within the ε-tubed as the final training dataset.

    • Sparse multikernel support vector regression machines trained by active learning

      2012, Expert Systems with Applications
      Citation Excerpt :

      Step 9: If the stopping criterion is not met, go to Step 3. In Zhao et al. (2010), limited training data set is also considered for multikernel support vector regression. Reducing samples strategy is utilised to select training samples for admissible functions.

    • New validation methods for improving standard and multi-parametric support vector regression training time

      2012, Expert Systems with Applications
      Citation Excerpt :

      In spite of this, the time for training a SVMr model can be very high due to the SVMr performance heavily depends on the choice of several hyper-parameters, necessary to define the optimization problem and the final SVMr model. There are different approaches focused on reducing this hard computation time of the SVMr model: in Guo and Zhang (2007) a method based on reducing the number of the samples included in the SVMr training is proposed, and in Zhao, Sun, and Zou (2010) a similar idea is applied to multi-parametric kernel SVMr. In Zhao and Sun (2010) a different methodology is applied, based on approximating the SVMr solution instead of solving the optimization problem in an exact way.

    • Heuristic sample reduction based support vector regression method

      2016, 2016 IEEE International Conference on Mechatronics and Automation, IEEE ICMA 2016
    • An activity improved robust adaptive normal SVR optimization algorithm

      2014, Journal of Computational Information Systems
    View all citing articles on Scopus
    View full text