Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
article

Feature subset selection using a new definition of classifiability

Published: 01 June 2003 Publication History
  • Get Citation Alerts
  • Abstract

    The performance of most practical classifiers improves when correlated or irrelevant features are removed. Machine based classification is thus often preceded by subset selection--a procedure which identifies relevant features of a high dimensional data set. At present, the most widely used subset selection technique is the so-called "wrapper" approach in which a search algorithm is used to identify candidate subsets and the actual classifier is used as a "black box" to evaluate the fitness of the subset. Fitness evaluation of the subset however requires cross-validation or other resampling based procedure for error estimation necessitating the construction of a large number of classifiers for each subset. This significant computational burden makes the wrapper approach impractical when a large number of features are present.In this paper, we present an approach to subset selection based on a novel definition of the classifiability of a given data. The classifiability measure we propose characterizes the relative ease with which some labeled data can be classified. We use this definition of classifiability to systematically add the feature which leads to the most increase in classifiability. The proposed approach does not require the construction of classifiers at each step and therefore does not suffer from as high a computational burden as a wrapper approach. Our results over several different data sets indicate that the results obtained are at least as good as that obtained with the wrapper approach.

    References

    [1]
    Almuallin, H., Dietterich, T.G., 1991. Learning with many irrelevant features. Proc. 9th Nat. Conf. on AI. pp. 547-552.
    [2]
    Cardie, C., 1993. Using decision trees to improve case based learning. Proc. 10th Int. Conf. on Machine Learning. pp. 25-32.
    [3]
    Dong, M., Kothari, R., 2001. Look-ahead based fuzzy decision tree induction. IEEE Trans. Fuzzy Syst. 9, 461-468.
    [4]
    Duda, R.O., Hart, P.E., 1973. Pattern Classification and Scene Analysis. John Wiley.
    [5]
    Efron, B., Tibshirani, T.J., 1993. An Introduction to the Bootstrap. Chapman and Hall, New York, NY.
    [6]
    Efron, B., Tibshirani, T.J., 1995. Cross-validation and the bootstrap: Estimating the error rate of a prediction rule. Technical Report, Stanford University.
    [7]
    Friedman, J.H., Rafsky, L.C., 1979. Multivariate generalizations of the Wald-Wolfowitz and Smirnof two-sample tests. Ann. Statist., 697-717.
    [8]
    Fukunaga, K., 1990. Introduction to Statistical Pattern Recognition. Academic Press, Boston, MA.
    [9]
    Goldberg, D., 1989. Genetic Algorithms in Search, Optimization, and Machine Learning. Addison-Wesley, Reading, MA.
    [10]
    Haralick, R.M., 1980. Statistical and structural approaches to texture. Proc. IEEE 67, 786-804.
    [11]
    Hartman, E.J., Keeler, J.D., Kowalski, J.M., 1990. Layered neural networks with Gaussian hidden units as universal approximators. Neural Comput. 2, 210-215.
    [12]
    Ho, T.K., Basu, M., 2002. Complexity measures of supervised classification problems. IEEE Trans. Pattern Anal. Machine Intell. 24, 289-300.
    [13]
    Hoekstra, A., Duin, R.P.W., 1996. On the nonlinearity of pattern classifiers. Proc. 13th Int. Conf. on Pattern Recognition, pp. 271-275.
    [14]
    Kira, K., Rendell, L.A., 1992. The feature selection problem: Traditional methods and a new algorithm. Proc. 10th Nat. Conf. on AI. pp. 129-134.
    [15]
    Kohavi, R., John, G.H., 1997. Wrappers for feature subset selection. Artifical Intell. 97, 273-324.
    [16]
    Langley, P., Iba, W., Thompson, K., 1992. An analysis of Bayesian classifiers. Proc. 10th Nat. Conf. on AI. pp. 223-228.
    [17]
    Narendra, P.M., Fukunaga, K., 1977. A branch and bound algorithm for feature subset selection. IEEE Trans. Comput. 26, 917-922.
    [18]
    Neter, J., Wasserman, W., Kutner, M.H., 1990. Applied Linear Statistical Models, third ed. Richard D. Irwin Inc, Englewood Cliffs, New Jersey.
    [19]
    Quinlan, J.R., 1986. Induction of decision trees. Machine Learning 1, 81-106.
    [20]
    Rao, A.R., 1990. A Taxonomy for Texture Description and Identification. Springer-Verlag, New York, NY.
    [21]
    Russell, S.J., Norvig, P., 1995. Artificial Intelligence: A Modern Approach. Prentice Hall, Englewood Cliff, NJ.
    [22]
    Smith, F.W., 1968. Pattern classifier design by linear programming. IEEE Trans. Comput. 17, 367-372.

    Cited By

    View all
    • (2022)Automatic Assessment of Quality of your Data for AIProceedings of the 5th Joint International Conference on Data Science & Management of Data (9th ACM IKDD CODS and 27th COMAD)10.1145/3493700.3493774(354-357)Online publication date: 8-Jan-2022
    • (2019)How Complex Is Your Classification Problem?ACM Computing Surveys10.1145/334771152:5(1-34)Online publication date: 13-Sep-2019
    • (2015)An automatic extraction method of the domains of competence for learning classifiers using data complexity measuresKnowledge and Information Systems10.1007/s10115-013-0700-442:1(147-180)Online publication date: 1-Jan-2015
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image Pattern Recognition Letters
    Pattern Recognition Letters  Volume 24, Issue 9-10
    01 June 2003
    501 pages

    Publisher

    Elsevier Science Inc.

    United States

    Publication History

    Published: 01 June 2003

    Author Tags

    1. classification
    2. dimensionality reduction
    3. feature selection

    Qualifiers

    • Article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 13 Aug 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2022)Automatic Assessment of Quality of your Data for AIProceedings of the 5th Joint International Conference on Data Science & Management of Data (9th ACM IKDD CODS and 27th COMAD)10.1145/3493700.3493774(354-357)Online publication date: 8-Jan-2022
    • (2019)How Complex Is Your Classification Problem?ACM Computing Surveys10.1145/334771152:5(1-34)Online publication date: 13-Sep-2019
    • (2015)An automatic extraction method of the domains of competence for learning classifiers using data complexity measuresKnowledge and Information Systems10.1007/s10115-013-0700-442:1(147-180)Online publication date: 1-Jan-2015
    • (2013)Predicting noise filtering efficacy with data complexity measures for nearest neighbor classificationPattern Recognition10.1016/j.patcog.2012.07.00946:1(355-364)Online publication date: 1-Jan-2013
    • (2013)Analysis of data complexity measures for classificationExpert Systems with Applications: An International Journal10.1016/j.eswa.2013.02.02540:12(4820-4831)Online publication date: 1-Sep-2013
    • (2012)Shared domains of competence of approximate learning models using measures of separability of classesInformation Sciences: an International Journal10.1016/j.ins.2011.09.022185:1(43-65)Online publication date: 1-Feb-2012
    • (2012)Linear separability and classification complexityExpert Systems with Applications: An International Journal10.1016/j.eswa.2012.01.09039:9(7796-7807)Online publication date: 1-Jul-2012
    • (2011)An accumulative points/votes based approach for feature selectionProceedings of the 16th Iberoamerican Congress conference on Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications10.1007/978-3-642-25085-9_47(399-408)Online publication date: 15-Nov-2011
    • (2009)Subspace clustering of images using ant colony optimisationProceedings of the 16th IEEE international conference on Image processing10.5555/1818719.1818801(229-232)Online publication date: 7-Nov-2009
    • (2008)Fast 3D reconstruction and recognitionProceedings of the 8th conference on Signal processing, computational geometry and artificial vision10.5555/1503734.1503737(15-21)Online publication date: 20-Aug-2008
    • Show More Cited By

    View Options

    View options

    Get Access

    Login options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media