Abstract
With the enormous growth of data, which exhibit sequentiality, it has become important to investigate the impact of embedded sequential information within the data. Sequential data are growing enormously, hence an efficient classification of sequential data is needed. k-Nearest Neighbor (kNN) has been used and proved to be an efficient classification technique for two-class problems. This paper uses sliding window approach to extract sub-sequences of various lengths and classification using kNN. We conducted experiments on DARPA 98 IDS dataset using various distance/similarity measures such as Jaccard similarity, Cosine similarity, Euclidian distance and Binary Weighted Cosine (BWC) measure. Our results demonstrate that sub-sequence information enhances kNN classification accuracy for sequential data, irrespective of the distance/similarity metric used.
An erratum to this chapter can be found at http://dx.doi.org/10.1007/11604655_67
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kumar, N.P., Rao, M.V., Krishna, P.R., Bapi, R.S. (2005). Using Sub-sequence Information with kNN for Classification of Sequential Data. In: Chakraborty, G. (eds) Distributed Computing and Internet Technology. ICDCIT 2005. Lecture Notes in Computer Science, vol 3816. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11604655_60
Download citation
DOI: https://doi.org/10.1007/11604655_60
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-30999-4
Online ISBN: 978-3-540-32429-4
eBook Packages: Computer ScienceComputer Science (R0)