Abstract
Feature selection is a very crucial step in data mining process. It aims to find the most important feature subset from a given feature set without degradation of classifying information. As for the traditional feature selection method, the number of candidate feature subsets created by algorithm in an iterative computational way is exponential in the size of the initial attribute set. And relevant algorithm occupies a lot of the system resources in time and space. In this paper, we study and develop a novel feature selection method and provide its mathematic principle, which is based on the factors of attributes contributing to target attribute and their maximum information divergence value (MIDV) to select small enough feature subset and improve the classification accuracy. And then the extensive experiment shows that our proposed method is very efficient in computational performance and scalability than traditional methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Mitchell, T.: Machine Learning. McGraw-Hill, New York (1997)
de Angelis, V., Felici, G., Mancinelli, G.: Feature selection for data mining. In: Data mining & Knowledge discovery based on rules reduction, pp. 227–237. Springer, Heidelberg (2006)
Langley, P.: Selection of relevant features in machine learning. In: Proceedings of the AAAI Fall Symposium on Relevance. AAAI Press, Menlo Park (1994)
Duda, R.O., Hart, P.E.: Pattern Classification, 2nd edn., pp. 90–101. Elsevier Science, USA (2003)
Fukunaga, Keinosuke: Introduction to Statistical Pattern Recognition, 2nd edn., pp. 489–503. Elsevier Academic Press (1999)
Ruiz, R., Aguilar–Ruiz, J.S.: Analysis of Feature Rankings for Classification, Spain
Abe, N., Kudo, M.: Entropy Criterion for Classifier-Independent Feature Selection, Japan
Abe, N., Kudo, M.: A Divergence Criterion for Classifier-Independent Feature Selection, Japan
Cang, S., Partridge, D.: Feature ranking and best feature subset using mutual information. Springer, London (2004)
Chizi, B., Maimon, O.: Data Mining & Knowledge Discovery Handbook. Springer Science, pp. 93–109 (2005)
Ian, H., Witten, I.H., Frank, E.: Data Mining: Practical Machine Learning Tools And Techniques, 2nd edn. Elsevier Inc., Amsterdam (2005)
Han, J., Kamber, M.: Data Mining: Concepts and Techniques, pp. 10–19. Morgan Kaufmann Publishers, San Francisco (2001)
Ye, N.: The Handbook of Data Mining, pp. 414–417. Lawrence Erlbaum Associates, Mahwah (2003)
Nixon, M.S., Aguado, A.S.: Feature Extraction and Image Processing. Elsevier Newnes, Amsterdam (2002)
Fukunaga, K.: Introduction to Statistical Pattern Recognition, 2nd edn., pp. 489–503. Elsevier Academic Press, Amsterdam (1999)
Marques de Sa, J.P.: Pattern Recognition Concepts Methods and Applications, pp. 65–69. Springer, Heidelberg (2001)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, Z., Shi, Y., Gao, G., Chai, Y. (2008). An Effective Feature Selection Method Using the Contribution Likelihood Ratio of Attributes for Classification. In: Ishikawa, Y., et al. Advanced Web and Network Technologies, and Applications. APWeb 2008. Lecture Notes in Computer Science, vol 4977. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-89376-9_16
Download citation
DOI: https://doi.org/10.1007/978-3-540-89376-9_16
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-89375-2
Online ISBN: 978-3-540-89376-9
eBook Packages: Computer ScienceComputer Science (R0)