Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

A method for combining mutual information and canonical correlation analysis

Published: 01 February 2012 Publication History

Abstract

Highlights We propose a hybrid measure of relevance based on MI and KCCA. Our measure PMI weighs more the samples with predictive powers. PMI effectively eliminates the samples with no predictive contribution. We show that PMI has improved feature detection capability. Feature selection is a critical step in many artificial intelligence and pattern recognition problems. Shannon's Mutual Information (MI) is a classical and widely used measure of dependence measure that serves as a good feature selection algorithm. However, as it is a measure of mutual information in average, under-sampled classes (rare events) can be overlooked by this measure, which can cause critical false negatives (missing a relevant feature very predictive of some rare but important classes). Shannon's mutual information requires a well sampled database, which is not typical of many fields of modern science (such as biomedical), in which there are limited number of samples to learn from, or at least, not all the classes of the target function (such as certain phenotypes in biomedical) are well-sampled. On the other hand, Kernel Canonical Correlation Analysis (KCCA) is a nonlinear correlation measure effectively used to detect independence but its use for feature selection or ranking is limited due to the fact that its formulation is not intended to measure the amount of information (entropy) of the dependence. In this paper, we propose a hybrid measure of relevance, Predictive Mutual Information (PMI) based on MI, which also accounts for predictability of signals from each other in its calculation as in KCCA. We show that PMI has more improved feature detection capability than MI, especially in catching suspicious coincidences that are rare but potentially important not only for experimental studies but also for building computational models. We demonstrate the usefulness of PMI, and superiority over MI, on both toy and real datasets.

References

[1]
Akaho, S. (2001). A kernel method for canonical correlation analysis. International Meeting of Psychometric Society (IMPS2001).
[2]
UCI machine learning repository. University of California, Department of Information and Computer Science, Irvine, CA.
[3]
Kernel independent component analysis. The Journal of Machine Learning Research. v3. 1-48.
[4]
Mutual information maximization: Models of cortical self-organization. Network. v7 i1. 7-31.
[5]
Mutual information and maximal correlation as measures of dependence. Annals of Mathematical Statistics. v33. 587-595.
[6]
Neural networks for pattern recognition. Oxford University Press, Oxford.
[7]
Estimating optimal transformations for multiple regression and correlation. Journal of the American Statistical Association. v80 i391. 580-598.
[8]
On the maximum correlation coefficient. Theory of Probability and Its Applications. v49. 132-138.
[9]
A tutorial on support vector machines for pattern recognition. Knowledge Discovery and Data Mining. v2. 1-43.
[10]
Bayesian bin distribution inference and mutual information. IEEE Transactions on Information Theory. v51 i11. 3766-3779.
[11]
SINBAD: A neocortical mechanism for discovering environmental variables and regularities hidden in sensory input. Biological Cybernetics. v90. 191-202.
[12]
Das statistische problem der correlation also variations und Eigenwert problem. Zeitschrift fur Angewandte Mathematik und Mechanik. v21. 364-379.
[13]
Gurrera, M. D. C. (2005). Construction of bivariate distributions and statistical dependence operations. Ph.D. dissertation, University of Barcelona, Spain.
[14]
Guvenir, H. A., Acar, B., Demiroz, G., & Cekin, A. (1997). A supervised machine learning algorithm for arrhythmia analysis. In Proceedings of the computers in cardiology conference. Lund, Sweden.
[15]
Canonical correlation analysis: An overview with application to learning methods. Neural Computation. v16. 2639-2664.
[16]
Relations between two sets of variates. Biometrika. v28. 321-377.
[17]
A comparison of methods for multi-class support vector machines. IEEE Transactions on Neural Networks. v13. 415-425.
[18]
The class imbalance problem: A systematic study. Intelligent Data Analysis. v6 i5. 429-450.
[19]
SINBAD automation of scientific discovery: From factor analysis to theory synthesis. Natural Computing. v3 i2. 207-233.
[20]
Feature selection and extraction using an unsupervised biologically-suggested approximation to Gebelein's maximal correlation. International Journal of Pattern Recognition and Artificial Intelligence. v24 i3. 337-358.
[21]
Input feature selection by mutual information based on Parzen window. IEEE Transactions on Pattern Analysis and Machine Intelligence. v24. 1667-1671.
[22]
Machine learning for science: State of the art and future prospects. Science. v293. 2051-2055.
[23]
On Gebelein's correlation coefficient. Statistics and Probability Letters. v69. 299-303.
[24]
Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence. v27. 1226-1238.
[25]
On measures of dependence. Acta Mathematica Academiae Scieniarum Hungaricae. v10. 441-451.
[26]
Sakar, C. O., & Kursun, O. (2010). A hybrid method for feature selection based on mutual information and canonical correlation analysis. 20th international conference on pattern recognition. Istanbul, Turkey.
[27]
Learning with kernels. MIT Press, Cambridge, MA.
[28]
A mathematical theory of communication. Bell System Technical Journal. v27. 379-423.
[29]
Tavallaee, M., Bagheri, E., Lu, W., & Ghorbani, A. A. (2009). A detailed analysis of the KDD CUP 99 data set. In Proceedings of the second IEEE international conference on computational intelligence for security and defense applications (Ottawa, Ontario, Canada, July 08-10, 2009) (pp. 53-58). IEEE Press, Piscataway, NJ.
[30]
The nature of statistical learning theory. Springer-Verlag, New York.
[31]
Training cost-sensitive neural networks with methods addressing the class imbalance problem. IEEE Transactions on Knowledge and Data Engineering. v18 i1. 63-77.

Cited By

View all
  1. A method for combining mutual information and canonical correlation analysis

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image Expert Systems with Applications: An International Journal
    Expert Systems with Applications: An International Journal  Volume 39, Issue 3
    February, 2012
    1661 pages

    Publisher

    Pergamon Press, Inc.

    United States

    Publication History

    Published: 01 February 2012

    Author Tags

    1. Canonical correlation
    2. Gebelein's Maximal Correlation
    3. Imbalanced datasets
    4. Mutual information
    5. Statistical dependence
    6. Suspicious coincidences

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 20 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all

    View Options

    View options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media