Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3077136.3080815acmconferencesArticle/Chapter ViewAbstractPublication PagesirConference Proceedingsconference-collections
research-article

Stacking Bagged and Boosted Forests for Effective Automated Classification

Published: 07 August 2017 Publication History

Abstract

Random Forest (RF) is one of the most successful strategies for automated classification tasks. Motivated by the RF success, recently proposed RF-based classification approaches leverage the central RF idea of aggregating a large number of low-correlated trees, which are inherently parallelizable and provide exceptional generalization capabilities. In this context, this work brings several new contributions to this line of research. First, we propose a new RF-based strategy (BERT) that applies the boosting technique in bags of extremely randomized trees. Second, we empirically demonstrate that this new strategy, as well as the recently proposed BROOF and LazyNN_RF classifiers do complement each other, motivating us to stack them to produce an even more effective classifier. Up to our knowledge, this is the first strategy to effectively combine the three main ensemble strategies: stacking, bagging (the cornerstone of RFs) and boosting. Finally, we exploit the efficient and unbiased stacking strategy based on out-of-bag (OOB) samples to considerably speedup the very costly training process of the stacking procedure. Our experiments in several datasets covering two high-dimensional and noisy domains of topic and sentiment classification provide strong evidence in favor of the benefits of our RF-based solutions. We show that BERT is among the top performers in the vast majority of analyzed cases, while retaining the unique benefits of RF classifiers (explainability, parallelization, easiness of parameterization). We also show that stacking only the recently proposed RF-based classifiers and BERT using our OOB-based strategy is not only significantly faster than recently proposed stacking strategies (up to six times) but also much more effective, with gains up to 21% and 17% on MacroF1 and MicroF1, respectively, over the best base method, and of 5% and 6% over a stacking of traditional methods, performing no worse than a complete stacking of methods at a much lower computational effort.

References

[1]
Ricardo A. Baeza-Yates and Berthier A. Ribeiro-Neto. 1999. Modern Information Retrieval. ACM Press / Addison-Wesley.
[2]
Yaxin Bi, David Bell, Hui Wang, Gongde Guo, and Jiwen Guan. 2007. COMBINING MULTIPLE CLASSIFIERS USING DEMPSTER'S RULE FOR TEXT CATEGORIZATION. Appl. Artif. Intell. 21, 3 (March 2007), 211--239.
[3]
Leo Breiman. 1996. Bagging Predictors. Mach. Learn. 24, 2 (Aug. 1996), 123--140.
[4]
A. Danesh, B. Moshiri, and O. Fatemi. 2007. Improve text classification accuracy based on classifier fusion methods. In Information Fusion, 2007 10th International Conference on. IEEE, 1--6.
[5]
Clebson C. A. de Sá, Marcos André Gonçalves, Daniel Xavier de Sousa, and Thiago Salles. 2016. Generalized BROOF-L2R: A General Framework for Learning to Rank Based on Boosting and Random Forests. In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, SIGIR 2016, Pisa, Italy, July 17-21, 2016. 95--104.
[6]
Yan-Shi Dong and Ke-Song Han. 2004. A comparison of several ensemble methods for text categorization. In Services Computing, 2004. (SCC 2004). Proceedings. 2004 IEEE International Conference on. IEEE, 419--422.
[7]
Yan-Shi Dong and Ke-Song Han. 2004. A Comparison of Several Ensemble Methods for Text Categorization. In IEEE International Conference on Services Computing. IEEE.
[8]
Manuel Fernández-Delgado, Eva Cernadas, Senén Barro, and Dinani Amorim. 2014. Do We Need Hundreds of Classifiers to Solve Real World Classification Problems? J. Mach. Learn. Res. 15, 1 (Jan. 2014), 3133--3181.
[9]
Yoav Freund and Robert E Schapire. 1997. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J. Comput. Syst. Sci. 55, 1 (Aug. 1997), 119--139.
[10]
Pierre Geurts, Damien Ernst, and Louis Wehenkel. 2006. Extremely randomized trees. Machine Learning 63, 1 (2006), 3--42.
[11]
Giorgio Giancinto and Fabio Roli. 2001. Design of Effective Neural Network Ensembles for Image Classification Purposes. IMAGE VISION AND COMPUTING JOURNAL 19 (2001), 699--707.
[12]
T. Hastie, R. Tibshirani, and J. H. Friedman. 2009. The Elements of Statistical Learning. Springer.
[13]
Monisha Kanakaraj and Ram Mohana Reddy Guddeti. 2015. Performance analysis of Ensemble methods on Twitter sentiment analysis using NLP techniques. In IEEE International Conference on Semantic Computing (ICSC). IEEE.
[14]
Aurangzeb Khan, Baharum Baharudin, Lam Hong Lee, Khairullah Khan, and Universiti Teknologi Petronas Tronoh. 2010. A Review of Machine Learning Algorithms for Text-Documents Classification. In Journal of Advances In Information Technology. Academy Publisher.
[15]
Ludmila I. Kuncheva, James C. Bezdek, and Robert P. W. Duin. 2001. Decision templates for multiple classifier fusion: an experimental comparison. Pattern Recognition 34 (2001), 299--314.
[16]
Ludmila I. Kuncheva and Christopher J. Whitaker. 2003. Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy. Mach. Learn. 51, 2 (May 2003), 181--207.
[17]
Christopher J. Merz. 1999. Using Correspondence Analysis to Combine Classifiers. Mach. Learn. 36, 1-2 (July 1999), 33--58.
[18]
Tien Thanh Nguyen, Thi Thu Thuy Nguyen, Xuan Cuong Pham, and Alan WeeChung Liew. 2016. A Novel Combining Classifier Method Based on Variational Inference. Pattern Recogn. 49, C (Jan. 2016), 198--212.
[19]
Aytuğ Onan, Serdar Korukoğlu, and Hasan Bulut. 2016. Ensemble of Keyword Extraction Methods and Classifiers in Text Classification. Expert Syst. Appl. 57, C (Sept. 2016), 232--247.
[20]
Gabriel Pui, Cheong Fung, Jeffrey Xu Yu, Haixun Wang, David W. Cheung, and Huan Liu. 2006. A Balanced Ensemble Approach to Weighting Classifiers for Text Classification. In ICDM '06. Sixth International Conference on Data Mining. IEEE.
[21]
Lior Rokach. 2009. Taxonomy for characterizing ensemble methods in classification tasks: A review and annotated bibliography. In Computational Statistics Data Analysis, In Press, Corrected Proof.
[22]
Thiago Salles, Marcos Gonçalves, and Leonardo Rocha. 2017. PhD Dissertation: Random Forest based Classifiers for Classification Tasks with Noisy Data. (2017). Federal University of Minas Gerais.
[23]
Thiago Salles, Marcos Gonçalves, Victor Rodrigues, and Leonardo Rocha. 2015. BROOF: Exploiting Out-of-Bag Errors, Boosting and Random Forests for Effective Automated Classification. In Proc. of the 38th International ACM SIGIR Conference on Inf. Retrieval. ACM, 353--362.
[24]
M. R. Segal. 2004. Machine Learning Benchmarks and Random Forest Regression. Technical Report. University of California.
[25]
Kai Ming Ting and Ian H. Witten. 1997. Stacking Bagged and Dagged Models. In Proceedings of the Fourteenth International Conference on Machine Learning (ICML '97). Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 367--375.
[26]
David Wolpert and William G. Macready. 1996. Combining Stacking With Bagging To Improve A Learning Algorithm. Technical Report.
[27]
David H. Wolpert. 1992. Stacked Generalization. Neural Networks 5 (1992), 241--259.

Cited By

View all
  • (2024)ChatGPT as a Text Annotation Tool to Evaluate Sentiment Analysis on South African Financial InstitutionsIEEE Access10.1109/ACCESS.2024.346437412(144017-144043)Online publication date: 2024
  • (2023)Machine learning predictive model for aspiration screening in hospitalized patients with acute strokeScientific Reports10.1038/s41598-023-34999-813:1Online publication date: 15-May-2023
  • (2023)Classification of Yoga Poses Using Integration of Deep Learning and Machine Learning TechniquesProceedings of International Conference on Recent Trends in Computing10.1007/978-981-19-8825-7_36(417-428)Online publication date: 21-Mar-2023
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
SIGIR '17: Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval
August 2017
1476 pages
ISBN:9781450350228
DOI:10.1145/3077136
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 07 August 2017

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. bagging
  2. boosting
  3. classification
  4. ensemble
  5. random forests
  6. stacking

Qualifiers

  • Research-article

Funding Sources

Conference

SIGIR '17
Sponsor:

Acceptance Rates

SIGIR '17 Paper Acceptance Rate 78 of 362 submissions, 22%;
Overall Acceptance Rate 792 of 3,983 submissions, 20%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)31
  • Downloads (Last 6 weeks)3
Reflects downloads up to 12 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2024)ChatGPT as a Text Annotation Tool to Evaluate Sentiment Analysis on South African Financial InstitutionsIEEE Access10.1109/ACCESS.2024.346437412(144017-144043)Online publication date: 2024
  • (2023)Machine learning predictive model for aspiration screening in hospitalized patients with acute strokeScientific Reports10.1038/s41598-023-34999-813:1Online publication date: 15-May-2023
  • (2023)Classification of Yoga Poses Using Integration of Deep Learning and Machine Learning TechniquesProceedings of International Conference on Recent Trends in Computing10.1007/978-981-19-8825-7_36(417-428)Online publication date: 21-Mar-2023
  • (2023)Parallel Extremely Randomized Decision Forests on Graphics Processors for Text ClassificationParallel Processing and Applied Mathematics10.1007/978-3-031-30442-2_7(83-94)Online publication date: 28-Apr-2023
  • (2020)A cross-entropy based stacking method in ensemble learningJournal of Intelligent & Fuzzy Systems10.3233/JIFS-20060039:3(4677-4688)Online publication date: 7-Oct-2020
  • (2020)A pragmatic approach to hierarchical categorization of research expertise in the presence of scarce informationInternational Journal on Digital Libraries10.1007/s00799-018-0260-z21:1(61-73)Online publication date: 1-Mar-2020
  • (2019)A Semantics Aware Random Forest for Text ClassificationProceedings of the 28th ACM International Conference on Information and Knowledge Management10.1145/3357384.3357891(1061-1070)Online publication date: 3-Nov-2019
  • (2019)Bridging Text Visualization and Mining: A Task-Driven SurveyIEEE Transactions on Visualization and Computer Graphics10.1109/TVCG.2018.283434125:7(2482-2504)Online publication date: 1-Jul-2019
  • (2019)Document Performance Prediction for Automatic Text ClassificationAdvances in Information Retrieval10.1007/978-3-030-15719-7_17(132-139)Online publication date: 7-Apr-2019
  • (2018)Semantically-Enhanced Topic ModelingProceedings of the 27th ACM International Conference on Information and Knowledge Management10.1145/3269206.3271797(893-902)Online publication date: 17-Oct-2018

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media