Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/2020408.2020565acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
poster

Common component analysis for multiple covariance matrices

Published: 21 August 2011 Publication History

Abstract

We consider the problem of finding a suitable common low dimensional subspace for accurately representing a given set of covariance matrices. With one covariance matrix, this is principal component analysis (PCA). For multiple covariance matrices, we term the problem Common Component Analysis (CCA). While CCA can be posed as a tensor decomposition problem, standard approaches to tensor decompositions have two critical issues: (i) tensor decomposition methods are iterative and rely on the initialization; (ii) for a given level of approximation error, it is difficult to choose a suitable low dimensionality. In this paper, we present a detailed analysis of CCA that yields an effective initialization and iterative algorithms for the problem. The proposed methodology has provable approximation guarantees w.r.t. the global maximum and also allows one to choose the dimensionality for a given level of approximation error. We also establish conditions under which the methodology will achieve the global maximum. We illustrate the effectiveness of the proposed method through extensive experiments on synthetic data as well as on two real stock market datasets, where major financial events can be visualized in low dimensions.

References

[1]
D. Achlioptas. Database-friendly random projections. In ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems, pages 274--281, 2001.
[2]
A. Agarwal, E. Hazan, S. Kale, and R. E. Schapire. Algorithms for portfolio management based on the Newton method. In Proceedings of the 23rd International Conference on Machine Learning (ICML), 2006.
[3]
T. Anderson. An Introduction to Multivariate Statistics, 3rd ed. John Wiley, 2003.
[4]
C. M. Bishop. Pattern Recognition and Machine Learning. Springer, 2007.
[5]
T. Bollerslev, J. Russell, and M. Watson. Volatility and Time Series Econometrics: Essays in Honor of Robert Engle. 2010.
[6]
D. Cai, X. He, and J. Han. Subspace learning based on tensor analysis. In Technical Report UIUCDCS-R-2005--2572, 2005.
[7]
T. M. Cover. Universal portfolios. Mathematical Finance, 1(1):1--29, 1991.
[8]
S. Dasgupta. Experiments with random projection. In Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence (UAI), pages 143--151, 2000.
[9]
C. Ding and J. Ye. Two-dimensional singular value decomposition (2DSVD) for 2D maps and images. In Proceedings of the 5th SIAM International Conference on Data Mining (SDM), pages 32--43, 2005.
[10]
R. Engle. Autoregressive conditional heteroskedasticity with estimates of the variance of UK inflation. Econometrica, 50:987--1008, 1982.
[11]
R. Engle. Dynamic conditional correlation: A simple class of multivariate generalized autoregressive conditional heteroscedasticity models. Journal of Business and Economic Statistics, 20:339--350, 2002.
[12]
B. N. Flury. Common principal components in k groups. Journal of American Statistical Association, 79(388):892--898, 1984.
[13]
B. N. Flury. Common Principal Components and Related Multivariate Models. John Wiley, 1988.
[14]
K. Fukunaga. Introduction to Statistical Pattern Recognition, 2nd edition. Academic Press, 1990.
[15]
G. H. Golub and C. V. Loan. Matrix Computations,3rd ed. Johns Hopkins University Press, 1996.
[16]
R. A. Harshman. Foundations of the PARAFAC procedure: Models and conditions for an explanatory multimodal factor analysis. UCLA Working Papers in Phonetics, 16:1--84, 1970.
[17]
R. A. Harshman. PARAFAC. Tutorial and applications. Chemometrics and Intelligent Laboratory Systems, 38(2):149--171, 1997.
[18]
D. Helmbold, R. Schapire, Y. Singer, and M. Warmuth. Online portfolio setection using multiplicative weights. Mathematical Finance, 8(4):325--347, 1998.
[19]
R. A. Horn and C. R. Johnson. Matrix Analysis. Cambridge University Press, 1985.
[20]
E. Kofidis and P. A. Regalia. On the best rank-1 approximation of higher-order supersymmetric tensors. SIAM Journal on Matrix Analysis and Applications, 23(3):863--884, 2000.
[21]
T. G. Kolda. Orthogonal tensor decompositions. SIAM Journal on Matrix Analysis and Applications, 23(1):243--255, 2001.
[22]
T. G. Kolda and B. W. Bader. Tensor decompositions and applications. SIAM Review, 51(3):455--500, 2009.
[23]
T. G. Kolda, B. W. Bader, and J. P. Kenny. Higher- order web link analysis using multilinear algebra. In Proceedings of the fifth IEEE International Conference on Data Mining (ICDM), pages 242--249, 2005.
[24]
P. M. Kroonenberg. Applied Multiway Data Analysis. Wiley, 2008.
[25]
P. M. Kroonenberg and J. de Leeuw. Principal component analysis of three-mode data by means of alternating least squares algorithms. Psychometrika, 45(1):69--97, 1980.
[26]
L. D. Lathauwer, B. D. Moor, J. Vandewalle, and J. V. A multilinear singular value decomposition. SIAM Journal on Matrix Analysis and Applications, 21(4):1253--1278, 2000.
[27]
L. D. Lathauwer, B. D. Moor, J. Vandewalle, and J. V. On the best rank-1 and rank-(r1,r2,. . .,rn) approximation of higher-order tensors. SIAM Journal on Matrix Analysis and Applications, 21(4):1324--1342, 2000.
[28]
D. D. Lee and H. S. Seung. Algorithms for non-negative matrix factorization. In Advances in Neural Information Processing Systems (NIPS), pages 556--562, 2001.
[29]
J. D. Leeuw and G. Michaildis. Majorization methods in statistics. Biostatistics, 9(3):432--441, 2008.
[30]
J. A. Patz, D. Campdell-Lendrum, T. Holloway, and J. A. Foley. Impact of regional climate change on human health. Nature, 438:310--317, 2005.
[31]
M. Pourahmadi, M. J. Daniels, and T. Park. Simultaneous modelling of the Cholesky decomposition of several covariance matrices. Journal of Multivariate Analysis, 98:568--587, 2006.
[32]
J. T. Scruggs and P. Glabadanidis. Risk premia and the dynamic covariance between stock and bond returns. Journal of finance and quantitative analysis, 38(2):295--316, 2003.
[33]
J. Sun, D. Tao, and C. Faloutsos. Beyond streams and graphs: Dynamic tensor analysis. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD), pages 374--383, 2006.
[34]
S. Tadjudin and D. A. Landgrebe. Covariance estimation with limited training samples. IEEE Transactions on Geoscience and Remote Sensing, 37(4), 1999.
[35]
L. R. Tucker. Some mathematical notes on three-mode factor analysis. Psychometrika, 31:279--311, 1966.
[36]
H. Wang, A. Banerjee, and D. Boley. Common component analysis for multiple covariance matrices. Technical Report, TR-10-017, University of Minnesota, Twin Cities, 2010. http://www.cs.umn.edu/tech_reports_upload/tr2010/10-017.pdf.
[37]
J. Ye. Generalized low rank approximations of matrices. Machine Learning Journal, 61:167--191, 2005.

Cited By

View all
  • (2022)Computational Tactics for Precision Cancer Network BiologyInternational Journal of Molecular Sciences10.3390/ijms23221439823:22(14398)Online publication date: 19-Nov-2022
  • (2021)Dimensionality Reduction of SPD Data Based on Riemannian Manifold Tangent Spaces and IsometryEntropy10.3390/e2309111723:9(1117)Online publication date: 27-Aug-2021
  • (2021)Multilinear Common Component Analysis via Kronecker Product RepresentationNeural Computation10.1162/neco_a_01425(1-28)Online publication date: 19-Jul-2021
  • Show More Cited By

Index Terms

  1. Common component analysis for multiple covariance matrices

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    KDD '11: Proceedings of the 17th ACM SIGKDD international conference on Knowledge discovery and data mining
    August 2011
    1446 pages
    ISBN:9781450308137
    DOI:10.1145/2020408
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 21 August 2011

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. PCA
    2. dimensionality reduction
    3. parafac
    4. tensor decompositions
    5. tucker

    Qualifiers

    • Poster

    Conference

    KDD '11
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)12
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 03 Oct 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2022)Computational Tactics for Precision Cancer Network BiologyInternational Journal of Molecular Sciences10.3390/ijms23221439823:22(14398)Online publication date: 19-Nov-2022
    • (2021)Dimensionality Reduction of SPD Data Based on Riemannian Manifold Tangent Spaces and IsometryEntropy10.3390/e2309111723:9(1117)Online publication date: 27-Aug-2021
    • (2021)Multilinear Common Component Analysis via Kronecker Product RepresentationNeural Computation10.1162/neco_a_01425(1-28)Online publication date: 19-Jul-2021
    • (2020)Penalized logistic regression using functional connectivity as covariates with an application to mild cognitive impairmentCommunications for Statistical Applications and Methods10.29220/CSAM.2020.27.6.60327:6(603-624)Online publication date: 30-Nov-2020
    • (2018)Dimensionality Reduction on SPD Manifolds: The Emergence of Geometry-Aware MethodsIEEE Transactions on Pattern Analysis and Machine Intelligence10.1109/TPAMI.2017.265504840:1(48-62)Online publication date: 1-Jan-2018
    • (2018)Estimation of Dynamic Sparse Connectivity Patterns From Resting State fMRIIEEE Transactions on Medical Imaging10.1109/TMI.2017.278655337:5(1224-1234)Online publication date: May-2018
    • (2018)Research on Prediction of Brain Functional Connectivity Based on Dynamic Bayesian Hierarchical Model2018 10th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC)10.1109/IHMSC.2018.00092(369-372)Online publication date: Aug-2018
    • (2018)Sparse common component analysis for multiple high-dimensional datasets via noncentered principal component analysisStatistical Papers10.1007/s00362-018-1045-6Online publication date: 22-Sep-2018
    • (2017)Variable habitat conditions drive species covariation in the human microbiotaPLOS Computational Biology10.1371/journal.pcbi.100543513:4(e1005435)Online publication date: 27-Apr-2017

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media