Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/238061.238074acmconferencesArticle/Chapter ViewAbstractPublication PagescoltConference Proceedingsconference-collections
Article
Free access

A randomized approximation of the MDL for stochastic models with hidden variables

Published: 01 January 1996 Publication History
First page of PDF

References

[1]
S. Amari, "Information geometry of the EM and em algorithms for neural networks," Neural Networks, vol 8, no. 9, pp.1379-1408, 1995.
[2]
B.S. Clarke and A.R. Barron. Informationtheoretic asymptotics of Bayes methods," IEEE Trans. Inform. Theory, IT-36, pp.453-471, 1990.
[3]
A.P. Dempster, N.M. Laird, and D.B. Rubin, "Maximum likelihood from incomplete data via the EM algorithm," J.R.Statist. Soc., B, 39, pp.1-38, 1977.
[4]
J. Dicbolt and C.P. Robert, "Estimation of finite mixture distributions through Bayesian sampling," J.R.Statist. Soc. B, vol 56, 2, pp.365-375, 1994.
[5]
B. Everitt and D. Hand, Finite Mizture Dzstrzbutions, London: Chapman and Hall, 1981.
[6]
A.E. Gelfand and A.F.M. Smith, "Sampling-based approach to calculating marginal densities," J. Am. Statist. Assoc., vol.85, pp.398-409, 1990.
[7]
S. Geman and D. Geman, "Stochastic relaxation, Gibbs distributions, and the Bayes restoration of images," IEEE Trans. on Pattern Analysis and Machzne Intelligence, PAMI-6, pp.721-741, 1984.
[8]
W.K. Hastings, "Monte Carlo sampling method using Markov chains and their applications," Biometrika, vol.57, pp.97-109, 1970.
[9]
D.P. Helmbold, R.E. Schapire, Y. Singer, and M.K. Warmuth, "A comparison of new and old algorithms for a mixture estimation problem," in Proc. of COLT'95, 1995, pp.69-78.
[10]
N. Metropolis, M.N. Rosenbluth, A.H. Rosenbluth, A.H. Teller, and E. Teller, "Equations of state calculations by fast computation machines," J. Chemical Physics, vol.21, pp.1087-1091, 1953.
[11]
E. Nummelin, General irreducible Markov chains and non-negative operators, Cambridge University Press, 1984.
[12]
M. Li and P. Vit~nyi, An Introduction to Kolmogorov Complexzty and Its Applicatwns. Springer Verlag, New York, 1993.
[13]
J. Rissanen, "Modeling by shortest data description,'' Automat~ca, vo1.14, pp.465-471, 1978.
[14]
J. Rissanen, "Minimum description length principle,'' IBM Res. Report, RJ 4131, 1983.
[15]
J. Rissanen, "Stochastic complexity," J. R. Star. Soc. B, vol.49, 3, pp.223-239, 1987.
[16]
J. Rissanen, Stochastic Complexity in Statistzcal Inquiry, World Scientific, Singapore, 1989.
[17]
J. Rissanen, "Fisher information and stochastic complexity," IEEE Trans. on Inform. Theory, IT- 42, I (1996), 40-47.
[18]
J. Rissanen, T. Speed, and B. Yu, "Density estimation by stochastic complexity," IEEE Trans. Inform. Theory, IT-38, pp.315-323, 1992.
[19]
J. Rissanen and B. Yu, "MDL learning," Progress in A utomatzons and Informatzon Systems, Springer Verlag, 1991.
[20]
G.O. Roberts and N.G. Polson, "On the geometric convergence of the Gibbs sampler," J.R.Statist. Soc. B, vol.56, 2, pp.377-384, 1994.
[21]
J. Rosenthal, "Minorization conditions and convergence rates for Markov chain Monte Carlo," Technical report No.9321, Univ. of Toronto, Dept. of Statistics, 1993.
[22]
J. Rosenthal, "Analysis of the Gibbs sampler for a model related to James-Stein estimators," Technical report No.9413, Univ. of Toronto, Dept. of Statistics, 1994.
[23]
M.A. Tanner and H.W. Wong, "The calculation of posterior distributions by data augmentation," Jr. American Statist. Assoc., vol.82, pp.528-550, 1987.
[24]
L. Tierney, "Exploring posterior distributions using Markov chains," in Proc. of 23rd Syrup. on the Interface, 1991, pp.563-570.
[25]
C.F.J. Wu, "On the convergence properties of the EM algorithm," Ann. Prob., vol 11, 95-103, 1983.
[26]
K. Yamanishi, "A learning criterion for stochastic rules," Machine Learning, vol.9, pp.165-203, 1992.
[27]
K. Yamanishi, "Probably almost discriminative learning," Machine Learning, vo1.18, pp.23-50, 1995.
[28]
K. Yamanishi, "A loss bound model for on-line stochastic prediction algorithms," Inform. Cornput., vol 119, 1, pp.39-54, 1995.
[29]
K. Yamanishi, "Randomized approximate aggregating strategies and their applications to prediction and discrimination," in Proc. of COLT'95, 1995, pp.83-90.
[30]
K. Yamanishi, "A decision-theoretic extension of stochastic complexity and its approximation to learning," submitted to IEEE Trans. Inform. Theory, 1995.
[31]
K. Yamanishi, "A randomized approximation of the minimum description length," submitted to IEEE Trans. Inform. Theory, 1995.

Cited By

View all
  • (2023)Information and CodingLearning with the Minimum Description Length Principle10.1007/978-981-99-1790-7_1(1-46)Online publication date: 15-Sep-2023
  • (2006)Minimum description length induction, Bayesianism, and Kolmogorov complexityIEEE Transactions on Information Theory10.1109/18.82580746:2(446-464)Online publication date: 1-Sep-2006
  • (1997)Document classification using a finite mixture modelProceedings of the 35th Annual Meeting of the Association for Computational Linguistics and Eighth Conference of the European Chapter of the Association for Computational Linguistics10.3115/976909.979623(39-47)Online publication date: 7-Jul-1997
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
COLT '96: Proceedings of the ninth annual conference on Computational learning theory
January 1996
344 pages
ISBN:0897918118
DOI:10.1145/238061
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 January 1996

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Conference

9COLT96
Sponsor:
9COLT96: 9th Annual Conference on Computational Learning Theory
June 28 - July 1, 1996
Desenzano del Garda, Italy

Acceptance Rates

Overall Acceptance Rate 35 of 71 submissions, 49%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)41
  • Downloads (Last 6 weeks)7
Reflects downloads up to 10 Sep 2024

Other Metrics

Citations

Cited By

View all
  • (2023)Information and CodingLearning with the Minimum Description Length Principle10.1007/978-981-99-1790-7_1(1-46)Online publication date: 15-Sep-2023
  • (2006)Minimum description length induction, Bayesianism, and Kolmogorov complexityIEEE Transactions on Information Theory10.1109/18.82580746:2(446-464)Online publication date: 1-Sep-2006
  • (1997)Document classification using a finite mixture modelProceedings of the 35th Annual Meeting of the Association for Computational Linguistics and Eighth Conference of the European Chapter of the Association for Computational Linguistics10.3115/976909.979623(39-47)Online publication date: 7-Jul-1997
  • (1997)Distributed cooperative Bayesian learning strategiesProceedings of the tenth annual conference on Computational learning theory10.1145/267460.267507(250-262)Online publication date: 1-Jul-1997
  • (1997)On prediction by data compressionProceedings of the 9th European Conference on Machine Learning10.1007/3-540-62858-4_69(14-30)Online publication date: 23-Apr-1997

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media