Abstract
In this chapter we are going to study the concept of support vector machines as developed by Vapnik and others. The concept was first proposed as an alternative to neural networks, when neural networks were not performing up to the grand expectations that they came with. SVM proposed a very targeted mathematical approach towards finding the optimal solution in case of classification or regression. We will first study the original SVM theory that tries to solve the problem of linear classification. Then, we will see how it can be further generalized for nonlinear problems with use of kernels and also how it is extended for solving the problems of regression. The theory of SVM proposed an elegant solution towards optimization and generalization and more importantly was extremely successful in getting results that neural network-based methods only hoped for at the time.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
A hyperplane is a generic term used to represent a linear plane in n dimensions. In one-dimensional space, it is a dot; in two-dimensional space, it is a line; and in three-dimensional space, it is a regular plane. The same concept can be extended to higher dimensions where the geometric manifold cannot be directly visualized, but can be mathematically modelled.
- 2.
Sometimes this is also called as kernel trick, although this is far more than a simple trick. A function needs to satisfy certain properties in order to be able called as kernel function. For more details on kernel functions, refer to [55].
References
Real World Datasets from Sklearn https://scikit-learn.org/stable/datasets/realworld.html
Vladimir N. Vapnik, The Nature of Statistical Learning Theory, 2nd edn. (Springer, New York, 1995).
V. N. Vapnik and A. Y. Lerner Pattern Recognition using Generazlied Portraits Automation and Remote Control, 24, 1963.
Olivier Chapelle, Jason Weston, Leon Bottou and Vladimir Vapnik, Vicinal Risk Minization, NIPS, 2000.
Vladimir Vapnik, Principles of Risk Minimization for Learning Theory, NIPS 1991.
Author information
Authors and Affiliations
Appendix
Appendix
New Lagrangian to be minimized is given as
Here, with the \(\max \) function, we are essentially ignoring the cases when there is error in the classification.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Joshi, A.V. (2023). Support Vector Machines. In: Machine Learning and Artificial Intelligence. Springer, Cham. https://doi.org/10.1007/978-3-031-12282-8_8
Download citation
DOI: https://doi.org/10.1007/978-3-031-12282-8_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-12281-1
Online ISBN: 978-3-031-12282-8
eBook Packages: Computer ScienceComputer Science (R0)