Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1390156.1390274acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
research-article

Data spectroscopy: learning mixture models using eigenspaces of convolution operators

Published: 05 July 2008 Publication History
  • Get Citation Alerts
  • Abstract

    In this paper we develop a spectral framework for estimating mixture distributions, specifically Gaussian mixture models. In physics, spectroscopy is often used for the identification of substances through their spectrum. Treating a kernel function K(x, y) as "light" and the sampled data as "substance", the spectrum of their interaction (eigenvalues and eigenvectors of the kernel matrix K) unveils certain aspects of the underlying parametric distribution p, such as the parameters of a Gaussian mixture. Our approach extends the intuitions and analyses underlying the existing spectral techniques, such as spectral clustering and Kernel Principal Components Analysis (KPCA).
    We construct algorithms to estimate parameters of Gaussian mixture models, including the number of mixture components, their means and covariance matrices, which are important in many practical applications. We provide a theoretical framework and show encouraging experimental results.

    References

    [1]
    M. Belkin, P. Niyogi, Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering, NIPS 2001.
    [2]
    S. Dasgupta, Learning mixtures of Gaussians, FOCS 1999.
    [3]
    A. Dempster, N. Laird, D. Rubin, Maximum-likelihood from incomplete data via the em algorithm, Journal of Royal Statistics Society, Ser. B, 39 (1997), pp. 1--38.
    [4]
    R. Kannan, H. Salmasian, S. Vempala, The spectral method for general mixture models, COLT 2005.
    [5]
    T. Kato, Perturbation Theory for Linear Operators, Springer-Verlag, 1966.
    [6]
    V. Koltchinskii and E. Giné, Random matrix approximation of spectra of integral operators, Bernoulli, 6 (2000), pp. 113--167.
    [7]
    S. MacEachern and P. Muller, Estimating mixture of Dirichlet process models, Journal of Computational and Graphical Statistics, 7 (1998), pp. 223--238.
    [8]
    A. Ng, M. Jordan, and Y. Weiss, On spectral clustering: Analysis and an algorithm, NIPS 2001.
    [9]
    K. Pearson, Contributions to the mathematical theory of evolution, Phil. Trans. Royal Soc., 185A, (1894), pp. 71--110.
    [10]
    R. Redner and H. Walker, Mixture densities, maximum likelihood and the em algorithm, SIAM Review, 26 (1984), pp. 195--239.
    [11]
    B. Schölkopf, A. J. Smola, and K.-R. Müller, Kernel principal component analysis, Advances in kernel methods: support vector learning, (1999), pp. 327--352.
    [12]
    C. Williams, M. Seeger, The effect of the input density distribution on kernel-based classifiers, ICML2000.
    [13]
    H. Zhu, C. Williams, R. Rohwer, and M. Morcinie, Gaussian regression and optimal finite dimensional linear models, in Neural networks and machine learning, C. Bishop, ed., 1998.

    Cited By

    View all

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    ICML '08: Proceedings of the 25th international conference on Machine learning
    July 2008
    1310 pages
    ISBN:9781605582054
    DOI:10.1145/1390156
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    • Pascal
    • University of Helsinki
    • Xerox
    • Federation of Finnish Learned Societies
    • Google Inc.
    • NSF
    • Machine Learning Journal/Springer
    • Microsoft Research: Microsoft Research
    • Intel: Intel
    • Yahoo!
    • Helsinki Institute for Information Technology
    • IBM: IBM

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 05 July 2008

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    ICML '08
    Sponsor:
    • Microsoft Research
    • Intel
    • IBM

    Acceptance Rates

    Overall Acceptance Rate 140 of 548 submissions, 26%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)4
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 12 Aug 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Graph signal interpolation and extrapolation over manifold of Gaussian mixtureSignal Processing10.1016/j.sigpro.2023.109308216(109308)Online publication date: Mar-2024
    • (2021)Interpreting uninterpretable predictors: kernel methods, Shtarkov solutions, and random forestsStatistical Theory and Related Fields10.1080/24754269.2021.1974157(1-19)Online publication date: 8-Sep-2021
    • (2021)Range Shifts Under Constant-Speed and Accelerated Climate WarmingBulletin of Mathematical Biology10.1007/s11538-021-00963-884:1Online publication date: 17-Nov-2021
    • (2020)On hyperparameter tuning in general clustering problemsProceedings of the 37th International Conference on Machine Learning10.5555/3524938.3525219(2996-3007)Online publication date: 13-Jul-2020
    • (2017)Kernel partial least squares for stationary dataThe Journal of Machine Learning Research10.5555/3122009.317686718:1(4447-4487)Online publication date: 1-Jan-2017
    • (2014)Distributed user Profiling via Spectral MethodsStochastic Systems10.1287/11-SSY0364:1(1-43)Online publication date: Jun-2014
    • (2012)Kernel current source density methodNeural Computation10.1162/NECO_a_0023624:2(541-575)Online publication date: 1-Feb-2012
    • (2012)Fast mining and forecasting of complex time-stamped eventsProceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining10.1145/2339530.2339577(271-279)Online publication date: 12-Aug-2012
    • (2012)Mixture weight influence on kernel entropy component analysis and semi-supervised learning using the Lasso2012 IEEE International Workshop on Machine Learning for Signal Processing10.1109/MLSP.2012.6349814(1-6)Online publication date: Sep-2012
    • (2012)Multiple Sample Data Spectroscopic Clustering of Large Datasets Using Nyström ExtensionJournal of Computational and Graphical Statistics10.1080/10618600.2012.67210421:2(338-360)Online publication date: 14-Jun-2012
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media