Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

A new optimizing parameter approach of LSSVM multiclass classification model

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

The parameter values of kernel function affect classification results to a certain extent. In the paper, a multiclass classification model based on improved least squares support vector machine (LSSVM) is presented. In the model, the non-sensitive loss function is replaced by quadratic loss function, and the inequality constraints are replaced by equality constraints. Consequently, quadratic programming problem is simplified as the problem of solving linear equation groups, and the SVM algorithm is realized by least squares method. When the LSSVM is used in multiclass classification, it is presented to choose parameter of kernel function on dynamic, which enhances preciseness rate of classification. The Fibonacci symmetry searching algorithm is simplified and improved. The changing rule of kernel function searching region and best shortening step is studied. The best multiclass classification results are obtained by means of synthesizing kernel function searching region and best shortening step. The simulation results show the validity of the model.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Cortes C, Vapnik V (1995) Support-vector network. Mach Learn 20(3):273–297

    MATH  Google Scholar 

  2. Vapnik V, Golowich SE, Smola A (1996) Support vector method for function approximation, regression estimation, and signal processing. Adv Neural Inf Process Syst 9:281–287

    Google Scholar 

  3. Chen S, Samingan AK, Hanzo L (2001) Support vector machine multiuser receiver for DS-CDMA signals in multipath channels. IEEE Trans Neural Netw 12(3):604–611

    Article  Google Scholar 

  4. Sebald DJ, Bucklew JA (2000) Support vector machine techniques for nonlinear equalization. IEEE Trans Signal Process 48(11):3217–3226

    Article  Google Scholar 

  5. Melgani F, Bruzzone L (2004) Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans Geosci Remote Sens 42(8):1778–1790

    Article  Google Scholar 

  6. Suykens JAK, Vandewalle J (1999) Least squares support vector machine classifiers. Neural Process Lett 9(3):293–300

    Article  MathSciNet  Google Scholar 

  7. Lukas L, Suykens JAK, Vandewalle J (2001) Least squares support vector machines classifiers: a multi two-spiral benchmark problem. In: Proceeding of the Indonesian Student Scientific Meeting. Manchester, United Kingdom, pp 289-292

  8. Suykens JAK, Lukas L, Vandewalle J (2000) Sparse least squares support vector machine classifiers. In: Proceedings of the European Symposium on Artificial Neural Networks. Bruges, Belgium, pp 37–42

  9. Bistarelli S, Codogent P, Hui HKC (2009) Solving finite domain constraint hierarchies by local consistency and tree search. J Exp Theor Artif Intell 21(4):233–257

    Article  MATH  Google Scholar 

  10. Chan KY, Kwong CK, Tsim YC (2010) Modeling and optimization of fluid dispensing for electronic packaging using neural fuzzy and genetic algorithms. Eng Appl Artif Intell 23:18–26

    Article  Google Scholar 

  11. Ge M, Du R, Zhang GC (2004) Fault diagnosis using support vector machine with an application in sheet metal stamping operation. Mech Syst Signal Process 18(1):143–159

    Article  Google Scholar 

  12. Hsu CW, Lin CJ (2002) A comparison of methods for multiclass support vector support vector machines. IEEE Trans Neural Netw 13(2):415–425

    Article  Google Scholar 

  13. Lendasse A, Simon G, Wertz V (2004) Fast bootstrap for least-square support vector machines. In: Proceedings of European Symposium on Artificial Neural Networks. Bruges, Belgium, pp 525–530

  14. Engel Y, Mannor S, Meir R (2004) The kernel recursive least-squares algorithm. IEEE Trans Signal Process 52(8):2275–2285

    Article  MathSciNet  Google Scholar 

  15. Qian SD (1990) Operational research. Tsinghua University Press, Beijing

    Google Scholar 

  16. Lin C J (2007) LIBSVM Data: Classification (Multi-class), http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass.html

Download references

Acknowledgments

This work is supported by China Postdoctoral Science Foundation (2005038515) and the Science Foundation of Hebei University of Science and Technology (XL2006081).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kui He Yang.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Yang, K.H., Zhao, L.L. A new optimizing parameter approach of LSSVM multiclass classification model. Neural Comput & Applic 21, 945–955 (2012). https://doi.org/10.1007/s00521-011-0673-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-011-0673-8

Keywords