Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1143844.1143911acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
Article

Using query-specific variance estimates to combine Bayesian classifiers

Published: 25 June 2006 Publication History

Abstract

Many of today's best classification results are obtained by combining the responses of a set of base classifiers to produce an answer for the query. This paper explores a novel "query specific" combination rule: After learning a set of simple belief network classifiers, we produce an answer to each query by combining their individual responses, using weights based inversely on their respective variances around their responses. These variances are based on the uncertainty of the network parameters, which in turn depend on the training datasample. In essence, this variance quantifies the base classifier's confidence of its response to this query. Our experimental results show that these "mixture-using-variance belief net classifiers" MUVS work effectively, especially when the base classifiers are learned using balanced bootstrap samples and when their results are combined using James-Stein shrinkage. We also found that our variance-based combination rule performed better than both bagging and AdaBoost, even on the set of base classifiers produced by AdaBoost itself. Finally, this framework is extremely efficient, as both the learning and the classification components require only straight-line code.

References

[1]
Bauer, E., & Kohavi, R. (1999). An empirical comparison of voting classification algorithms: Bagging, boosting, and variants. Machine Learning, 36.
[2]
Breiman, L. (1996). Bagging predictors. Machine Learning, 24.
[3]
Cooper, G., & Herskovits, E. (1992). A Bayesian method for the induction of probabilistic networks from data. Machine Learning, 9.
[4]
Duda, R., Hart, P., & Stork, D. (2002). Pattern classification, 2nd ed. Wiley.
[5]
Freund, Y., & Schapire, R. (1997). A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences, 55.
[6]
Friedman, N., Geiger, D., & Goldszmidt, M. (1997). Bayesian network classifiers. Machine Learning, 29.
[7]
Hastie, T., Tibshirani, R., & Friedman, J. (2002). The elements of statistical learning. Springer.
[8]
Heckerman, D. (1998). A tutorial on learning with Bayesian networks. Learning in Graphical Models.
[9]
Holte, R. (1993). Very simple classification rules perform well on most commonly used datasets. Machine Learning, 3.
[10]
Inza, I., Larranaga, P., Lozano, J., & Pena, J. (2005). Special issue of Machine Learning Journal: Probabilistic graphical models for classification, vol. 59.
[11]
Jacobs, R., Jordan, M., Nowlan, S., & Hinton, G. (1991). Adaptive mixtures of local experts. Neural Computation.
[12]
Krogh, A., & Vedelsby, J. (1995). Neural network ensembles, cross validation, and active learning. NIPS (pp. 231--238).
[13]
Lehmann, E., & Casella, G. (2003). Theory of point estimation. Springer.
[14]
Mitchell, T. (1997). Machine learning. McGraw-Hill.
[15]
Opitz, D., & Maclin, R. (1999). Popular ensemble methods: An empirical study. Journal of Artificial Intelligence Research, 11.
[16]
Pearl, J. (1988). Probabilistic reasoning in intelligent systems. Morgan Kaufmann.
[17]
Perrone, M., & Cooper, L. (1993). When networks disagree: ensemble method for neural networks. Artificial Neural Networks for Speech and Vision.
[18]
Sollich, P., & Krogh, A. (1996). Learning with ensembles: How overfitting can be useful. NIPS.
[19]
Taniguchi, M., & Tresp, V. (1997). Averaging regularized estimators. Neural Computation, 9.
[20]
UCI (2006). http://www.ics.uci.edu/~mlearn/.
[21]
Van Allen, T., Greiner, R., & Hooper, P. (2001). Bayesian error-bars for belief net inference. UAI.
[22]
Web (A). http://www.cs.ualberta.ca/~greiner/MUV.
[23]
Web (B). http://www.cs.ualberta.ca/~greiner/BNvar.

Cited By

View all
  • (2012)Combining Classifiers and Learning Mixture-of-ExpertsMachine Learning10.4018/978-1-60960-818-7.ch209(243-252)Online publication date: 2012
  • (2012)Best linear unbiased estimator for Kalman filter based left ventricle tracking in 3D+T echocardiography2012 IEEE Workshop on Mathematical Methods in Biomedical Image Analysis10.1109/MMBIA.2012.6164741(201-208)Online publication date: Jan-2012
  • (2011)Maximum likelihood and James-Stein edge estimators for left ventricle tracking in 3D echocardiographyProceedings of the Second international conference on Machine learning in medical imaging10.5555/2046063.2046069(43-50)Online publication date: 18-Sep-2011
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICML '06: Proceedings of the 23rd international conference on Machine learning
June 2006
1154 pages
ISBN:1595933832
DOI:10.1145/1143844
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 25 June 2006

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Acceptance Rates

ICML '06 Paper Acceptance Rate 140 of 548 submissions, 26%;
Overall Acceptance Rate 140 of 548 submissions, 26%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)2
  • Downloads (Last 6 weeks)0
Reflects downloads up to 16 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2012)Combining Classifiers and Learning Mixture-of-ExpertsMachine Learning10.4018/978-1-60960-818-7.ch209(243-252)Online publication date: 2012
  • (2012)Best linear unbiased estimator for Kalman filter based left ventricle tracking in 3D+T echocardiography2012 IEEE Workshop on Mathematical Methods in Biomedical Image Analysis10.1109/MMBIA.2012.6164741(201-208)Online publication date: Jan-2012
  • (2011)Maximum likelihood and James-Stein edge estimators for left ventricle tracking in 3D echocardiographyProceedings of the Second international conference on Machine learning in medical imaging10.5555/2046063.2046069(43-50)Online publication date: 18-Sep-2011
  • (2010)Learning to combine discriminative classifiersProceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining10.1145/1835804.1835899(743-752)Online publication date: 25-Jul-2010
  • (2009)Improved mean and variance approximations for belief net responses via network doublingProceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence10.5555/1795114.1795142(232-239)Online publication date: 18-Jun-2009
  • (2008)Quantifying the uncertainty of a belief net responseArtificial Intelligence10.1016/j.artint.2007.09.004172:4-5(483-513)Online publication date: 1-Mar-2008
  • (2007)Estimation and use of uncertainty in pseudo-relevance feedbackProceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval10.1145/1277741.1277795(303-310)Online publication date: 23-Jul-2007
  • (2007)How Relevant is Game Theory to Intelligent Agent Technology?Proceedings of the IEEE/WIC/ACM International Conference on Web Intelligence10.1109/WI.2007.65Online publication date: 2-Nov-2007
  • (2007)Neighborhood-Based Local SensitivityProceedings of the 18th European conference on Machine Learning10.1007/978-3-540-74958-5_7(30-41)Online publication date: 17-Sep-2007

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media