Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2908812.2908823acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Fast Bi-Objective Feature Selection Using Entropy Measures and Bayesian Inference

Published: 20 July 2016 Publication History

Abstract

The entropy measures have been used in feature selection for decades, and showed competitive performance. In general, the problem aims at minimizing the conditional entropy of the class label on the selected features. However, the generalization of the entropy measures has been neglected in literature. Specifically, the use of conditional entropy has two critical issues. First, the empirical conditional distribution of the class label may have a low confidence and thus is unreliable. Second, there may not be enough training instances for the selected features, and it is highly likely to encounter new examples in the test set. To address these issues, a bi-objective optimization model with a modified entropy measure called the Bayesian entropy is proposed. This model considers the confidence of the optimized conditional entropy value as well as the conditional entropy value itself. As a result, it produces multiple feature subsets with different trade-offs between the entropy value and its confidence. The experimental results demonstrate that by solving the proposed optimization model with the new entropy measure, the number of features can be dramatically reduced within a much shorter time than the existing algorithms. Furthermore, similar or even better classification accuracy was achieved for most test problems.

References

[1]
K. Bache and M. Lichman. Uci machine learning repository, 2013.
[2]
L. Bobrowski. Feature selection based on some homogeneity coefficient. In 9th International Conference on Pattern Recognition, pages 544--546. IEEE, 1988.
[3]
L. Cervante, B. Xue, L. Shang, and M. Zhang. A multi-objective feature selection approach based on binary PSO and rough set theory. In 13th European Conference on Evolutionary Computation in Combinatorial Optimization (EvoCOP), volume 7832 of Lecture Notes in Computer Science, pages 25--36. Springer, 2013.
[4]
L. Cervante, B. Xue, M. Zhang, and L. Shang. Binary particle swarm optimisation for feature selection: A filter based approach. In IEEE Congress on Evolutionary Computation (CEC'12), pages 881--888, 2012.
[5]
M. Dash and H. Liu. Feature selection for classification. Intelligent Data Analysis, 1(4):131--156, 1997.
[6]
K. Deb, S. Agrawal, A. Pratap, and T. Meyarivan. A fast and elitist multi-objective genetic algorithm: NSGA-II. IEEE Transactions on Evolutionary Computation, 6(2):182--197, 2002.
[7]
R. Diao and Q. Shen. Feature selection with harmony search. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 42(6):1509--1523, 2012.
[8]
P. Estévez, M. Tesmer, C. Perez, J. M. Zurada, et al. Normalized mutual information feature selection. IEEE Transactions on Neural Networks, 20(2):189--201, 2009.
[9]
I. Guyon and A. Elisseeff. An introduction to variable and feature selection. The Journal of Machine Learning Research, 3:1157--1182, 2003.
[10]
E. Hancer, B. Xue, D. Karaboga, and M. Zhang. A binary ABC algorithm based on advanced similarity scheme for feature selection. Applied Soft Computing, 36:334 -- 348, 2015.
[11]
M. M. Kabir, M. Shahjahan, and K. Murase. A new hybrid ant colony optimization algorithm for feature selection. Expert Systems with Applications, 39(3):3747--3763, 2012.
[12]
S. S. Kannan and N. Ramaraj. A novel hybrid feature selection via symmetrical uncertainty ranking based local memetic search algorithm. Knowledge-Based Systems, 23(6):580--585, 2010.
[13]
R. Kohavi and G. H. John. Wrappers for feature subset selection. Artificial intelligence, 97(1):273--324, 1997.
[14]
P. P. Kundu and S. Mitra. Multi-objective optimization of shared nearest neighbor similarity for feature selection. Applied Soft Computing, 37:751 -- 762, 2015.
[15]
H. Liu, H. Motoda, R. Setiono, and Z. Zhao. Feature selection: An ever evolving frontier in data mining. In Feature Selection for Data Mining, volume 10 of JMLR Proceedings, pages 4--13. JMLR.org, 2010.
[16]
H. Liu and L. Yu. Toward integrating feature selection algorithms for classification and clustering. IEEE Transactions on Knowledge and Data Engineering, 17(4):491--502, 2005.
[17]
H. B. Nguyen, B. Xue, I. Liu, and M. Zhang. Filter based backward elimination in wrapper based PSO for feature selection in classification. In IEEE Congress on Evolutionary Computation (CEC), pages 3111 -- 3118, 2014.
[18]
H. Peng, F. Long, and C. Ding. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(8):1226--1238, 2005.
[19]
C. Shannon and W. Weaver. The Mathematical Theory of Communication. Urbana:The University of Illinois Press, 1949.
[20]
S. S. Shreem, S. Abdullah, and M. Z. A. Nazri. Hybrid feature selection algorithm using symmetrical uncertainty and a harmony search algorithm. International Journal of Systems Science, (0):1--18, 2014.
[21]
N. Spolaôr, A. Lorena, and H. Lee. Multi-objective genetic algorithm evaluation in feature selection. In Evolutionary Multi-Criterion Optimization, volume 6576 of Lecture Notes in Computer Science, pages 462--476. Springer Berlin Heidelberg, 2011.
[22]
R. W. Swiniarski and A. Skowron. Rough set methods in feature selection and recognition. Pattern recognition letters, 24(6):833--849, 2003.
[23]
A. Unler and R. B. C. Alper Murat. mr2PSO: A maximum relevance minimum redundancy feature selection method based on swarm intelligence for support vector machine classification. Information Sciences, 20:4625--4641, 2011.
[24]
B. Xue, L. Cervante, L. Shang, W. N. Browne, and M. Zhang. Binary PSO and rough set theory for feature selection: A multi-objective filter based approach. International Journal of Computational Intelligence and Applications, 13(02):1450009:1--34, 2014.
[25]
B. Xue, M. Zhang, W. Browne, and X. Yao. A survey on evolutionary computation approaches to feature selection. IEEE Transactions on Evolutionary Computation, (99):1--1. 2015.
[26]
B. Xue, M. Zhang, and W. N. Browne. Particle swarm optimization for feature selection in classification: A multi-objective approach. IEEE Transactions on Cybernetics, 43(6):1656--1671, 2013.
[27]
L. Yu and H. Liu. Feature selection for high-dimensional data: A fast correlation-based filter solution. In ICML, volume 3, pages 856--863, 2003.
[28]
Z. Zhu, S. Jia, and Z. Ji. Towards a memetic feature selection paradigm {application notes}. IEEE Computational Intelligence Magazine, 5(2):41--53, 2010.

Cited By

View all
  • (2022)Attribute Selection Based Genetic Network Programming for Intrusion Detection SystemJournal of Advanced Computational Intelligence and Intelligent Informatics10.20965/jaciii.2022.p067126:5(671-683)Online publication date: 20-Sep-2022
  • (2016)Dimension reduction in classification using particle swarm optimisation and statistical variable grouping information2016 IEEE Symposium Series on Computational Intelligence (SSCI)10.1109/SSCI.2016.7850126(1-8)Online publication date: Dec-2016

Index Terms

  1. Fast Bi-Objective Feature Selection Using Entropy Measures and Bayesian Inference

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    GECCO '16: Proceedings of the Genetic and Evolutionary Computation Conference 2016
    July 2016
    1196 pages
    ISBN:9781450342063
    DOI:10.1145/2908812
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 20 July 2016

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. feature selection
    2. generalization
    3. multi-objective computation

    Qualifiers

    • Research-article

    Conference

    GECCO '16
    Sponsor:
    GECCO '16: Genetic and Evolutionary Computation Conference
    July 20 - 24, 2016
    Colorado, Denver, USA

    Acceptance Rates

    GECCO '16 Paper Acceptance Rate 137 of 381 submissions, 36%;
    Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 04 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2022)Attribute Selection Based Genetic Network Programming for Intrusion Detection SystemJournal of Advanced Computational Intelligence and Intelligent Informatics10.20965/jaciii.2022.p067126:5(671-683)Online publication date: 20-Sep-2022
    • (2016)Dimension reduction in classification using particle swarm optimisation and statistical variable grouping information2016 IEEE Symposium Series on Computational Intelligence (SSCI)10.1109/SSCI.2016.7850126(1-8)Online publication date: Dec-2016

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media