Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
research-article

Hierarchical Pruning of Deep Ensembles with Focal Diversity

Published: 16 January 2024 Publication History

Abstract

Deep neural network ensembles combine the wisdom of multiple deep neural networks to improve the generalizability and robustness over individual networks. It has gained increasing popularity to study and apply deep ensemble techniques in the deep learning community. Some mission-critical applications utilize a large number of deep neural networks to form deep ensembles to achieve desired accuracy and resilience, which introduces high time and space costs for ensemble execution. However, it still remains a critical challenge whether a small subset of the entire deep ensemble can achieve the same or better generalizability and how to effectively identify these small deep ensembles for improving the space and time efficiency of ensemble execution. This article presents a novel deep ensemble pruning approach, which can efficiently identify smaller deep ensembles and provide higher ensemble accuracy than the entire deep ensemble of a large number of member networks. Our hierarchical ensemble pruning approach (HQ) leverages three novel ensemble pruning techniques. First, we show that the focal ensemble diversity metrics can accurately capture the complementary capacity of the member networks of an ensemble team, which can guide ensemble pruning. Second, we design a focal ensemble diversity based hierarchical pruning approach, which will iteratively find high quality deep ensembles with low cost and high accuracy. Third, we develop a focal diversity consensus method to integrate multiple focal diversity metrics to refine ensemble pruning results, where smaller deep ensembles can be effectively identified to offer high accuracy, high robustness and high ensemble execution efficiency. Evaluated using popular benchmark datasets, we demonstrate that the proposed hierarchical ensemble pruning approach can effectively identify high quality deep ensembles with better classification generalizability while being more time and space efficient in ensemble decision making. We have released the source codes on GitHub at https://github.com/git-disl/HQ-Ensemble.

References

[1]
Lerina Aversano, Mario Luca Bernardi, Marta Cimitile, and Riccardo Pecori. 2021. Deep neural networks ensemble to detect COVID-19 from CT scans. Pattern Recognition 120, 12 (2021), 108135. DOI:
[2]
Robert E. Banfield, Lawrence O. Hall, Kevin W. Bowyer, and W. Philip Kegelmeyer. 2005. Ensemble diversity measures and their application to thinning. Information Fusion 6, 1 (2005), 49–62. DOI:Diversity in Multiple Classifier Systems.
[3]
Yijun Bian, Yijun Wang, Yaqiang Yao, and Huanhuan Chen. 2020. Ensemble pruning based on objection maximization with a general distributed framework. IEEE Transactions on Neural Networks and Learning Systems 31, 9 (2020), 3766–3774. DOI:
[4]
Leo Breiman. 1996. Bagging predictors. Machine Learning 24, 2 (1996), 123–140.
[5]
Leo Breiman. 2001. Random forests. In Proceedings of the Machine Learning. 5–32.
[6]
Leo Breiman. 1998. Arcing classifier (with discussion and a rejoinder by the author). The Annals of Statistics 26, 3 (1998), 801–849.
[7]
Rich Caruana, Alexandru Niculescu-Mizil, Geoff Crew, and Alex Ksikes. 2004. Ensemble selection from libraries of models. In Proceedings of the 21st International Conference on Machine Learning.Association for Computing Machinery, New York, NY, 18. DOI:
[8]
Ka-Ho Chow and Ling Liu. 2021. Robust object detection fusion against deception. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining.Association for Computing Machinery, New York, NY, 2703–2713. DOI:
[9]
Ka-Ho Chow, Wenqi Wei, Yanzhao Wu, and Ling Liu. 2019. Denoising and verification cross-layer ensemble against black-box adversarial attacks. In Proceedings of the 2019 IEEE International Conference on Big Data. 1282–1291. DOI:
[10]
Stanislav Fort, Huiyi Hu, and Balaji Lakshminarayanan. 2020. Deep Ensembles: A Loss Landscape Perspective. arXiv:1912.02757. Retrieved from https://arxiv.org/abs/1912.02757
[11]
Shenda Hong, Yanbo Xu, Alind Khare, Satria Priambada, Kevin Maher, Alaa Aljiffry, Jimeng Sun, and Alexey Tumanov. 2020. HOLMES: Health online model ensemble serving for deep learning models in intensive care units. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.Association for Computing Machinery, New York, NY, 1614–1624. DOI:
[12]
Chia-Yu Hsu and Ju-Chien Chien. 2022. Ensemble convolutional neural networks with weighted majority for wafer bin map pattern classification. Journal of Intelligent Manufacturing 33, 3 (2022), 831–844. DOI:
[13]
Cheng Ju, Aurélien Bibaut, and Mark Laan. 2018. The relative performance of ensemble methods with deep convolutional neural networks for image classification. Journal of Applied Statistics 45, 15(2018), 2800–2818.
[14]
Myeongso Kim, Minyoung Lee, Minjeong An, and Hongchul Lee. 2020. Effective automatic defect classification process based on CNN with stacking ensemble model for TFT-LCD panel. Journal of Intelligent Manufacturing 31, 5 (2020), 1165–1174. DOI:
[15]
Ron Kohavi and David Wolpert. 1996. Bias plus variance decomposition for zero-one loss functions. In Proceedings of the 13th International Conference on International Conference on Machine Learning.Morgan Kaufmann Publishers Inc., San Francisco, CA, 275–283.
[16]
Alex Krizhevsky and Geoffrey Hinton. 2009. Learning Multiple Layers of Features from Tiny Images. Technical Report. University of Toronto.
[17]
Ludmila I. Kuncheva and Christopher J. Whitaker. 2003. Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Machine Learning. 51, 2(2003), 181–207. DOI:
[18]
A. Lazarevic and Z. Obradovic. 2001. Effective pruning of neural network classifier ensembles. In IJCNN’01. International Joint Conference on Neural Networks. Proceedings (Cat. No.01CH37222)796–801.
[19]
Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner. 1998. Gradient-based learning applied to document recognition. Proc. IEEE 86, 11(1998), 2278–2324.
[20]
L. Liu, W. Wei, K. Chow, M. Loper, E. Gursoy, S. Truex, and Y. Wu. 2019. Deep neural network ensembles against deception: Ensemble diversity, accuracy and robustness. In Proceedings of the 2019 IEEE 16th International Conference on Mobile Ad Hoc and Sensor Systems. 274–282.
[21]
Tie-Yan Liu. 2011. Learning to Rank for Information Retrieval. Springer.
[22]
Yang Liu, Zhiyuan Liu, Cheng Lyu, and Jieping Ye. 2020. Attention-based deep ensemble net for large-scale online taxi-hailing demand prediction. IEEE Transactions on Intelligent Transportation Systems 21, 11 (2020), 4798–4807. DOI:
[23]
Qing Lu and Lise Getoor. 2003. Link-based classification. In Proceedings of the 20th International Conference on International Conference on Machine Learning.AAAI Press, 496–503.
[24]
G. Martínez-Muñoz, D. Hernández-Lobato, and A. Suárez. 2009. An analysis of ensemble pruning techniques based on ordered aggregation. IEEE Transactions on Pattern Analysis and Machine Intelligence 31, 2 (2009), 245–259.
[25]
Mary L. McHugh. 2012. Interrater reliability: the kappa statistic. Biochemia medica 22, 3 (2012), 276—282. Retrieved from https://europepmc.org/articles/PMC3900052
[26]
Kyushik Min, Dongchan Kim, Jongwon Park, and Kunsoo Huh. 2019. RNN-based path prediction of obstacle vehicles with deep ensemble. IEEE Transactions on Vehicular Technology 68, 10 (2019), 10252–10256. DOI:
[27]
D. Partridge and W. Krzanowski. 1997. Software diversity: Practical statistics for its measurement and exploitation. Information and Software Technology 39, 10 (1997), 707–717. DOI:
[28]
Remigijus Paulavičius and Julius Žilinskas. 2006. Analysis of different norms and corresponding Lipschitz constants for global optimization. Ukio Technologinis ir Ekonominis Vystymas 12, 4 (2006), 301–306.
[29]
Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. 2015. ImageNet large scale visual recognition challenge. International Journal of Computer Vision 115, 3 (2015), 211–252. DOI:
[30]
David B. Skalak. 1996. The sources of increased accuracy for two proposed boosting algorithms. In Proceedings of the American Association for Arti Intelligence, AAAI-96, Integrating Multiple Learned Models Workshop. 120–125.
[31]
E. Ke Tang, Ponnuthurai N Suganthan, and Xin Yao. 2006. An analysis of diversity measures. Machine Learning 65, 1 (2006), 247–271.
[32]
Grigorios Tsoumakas, Ioannis Partalas, and Ioannis Vlahavas. 2009. An Ensemble Pruning Primer. Springer, Berlin, 1–13.
[33]
KAGAN TUMER and JOYDEEP GHOSH. 1996. Error correlation and error reduction in ensemble classifiers. Connection Science 8, 3-4 (1996), 385–404. DOI:
[34]
Wenqi Wei and Ling Liu. 2021. Robust deep learning ensemble against deception. IEEE Transactions on Dependable and Secure Computing 18, 4 (2021), 1513–1527. DOI:
[35]
W. Wei, L. Liu, M. Loper, K. Chow, E. Gursoy, S. Truex, and Y. Wu. 2020. Cross-layer strategic ensemble defense against adversarial examples. In Proceedings of the 2020 International Conference on Computing, Networking and Communications. 456–460.
[36]
Yanzhao Wu, Ka-Ho Chow, Wenqi Wei, and Ling Liu. 2023. Exploring model learning heterogeneity for boosting ensemble robustness. In Proceedings of the 2023 IEEE International Conference on Data Mining.
[37]
Yanzhao Wu and Ling Liu. 2021. Boosting deep ensemble performance with hierarchical pruning. In Proceedings of the 2021 IEEE International Conference on Data Mining. 1433–1438. DOI:
[38]
Yanzhao Wu, Ling Liu, Zhongwei Xie, Juhyun Bae, Ka-Ho Chow, and Wenqi Wei. 2020. Promoting high diversity ensemble learning with ensemblebench. In Proceedings of the 2020 IEEE 2nd International Conference on Cognitive Machine Intelligence. 208–217. DOI:
[39]
Yanzhao Wu, Ling Liu, Zhongwei Xie, Ka-Ho Chow, and Wenqi Wei. 2021. Boosting ensemble accuracy by revisiting ensemble diversity metrics. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 16464–16472. DOI:
[40]
Xu-Cheng Yin, Chun Yang, and Hong-Wei Hao. 2014. Learning to Diversify via Weighted Kernels for Classifier Ensemble. arXiv:1406.1167. Retrieved from https://arxiv.org/abs/1406.1167

Cited By

View all
  • (2024)Effective Diversity Optimizations for High Accuracy Deep Ensembles2024 IEEE 6th International Conference on Cognitive Machine Intelligence (CogMI)10.1109/CogMI62246.2024.00044(278-287)Online publication date: 28-Oct-2024

Index Terms

  1. Hierarchical Pruning of Deep Ensembles with Focal Diversity

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Intelligent Systems and Technology
    ACM Transactions on Intelligent Systems and Technology  Volume 15, Issue 1
    February 2024
    533 pages
    EISSN:2157-6912
    DOI:10.1145/3613503
    • Editor:
    • Huan Liu
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 16 January 2024
    Online AM: 17 November 2023
    Accepted: 19 October 2023
    Revised: 17 May 2023
    Received: 18 July 2022
    Published in TIST Volume 15, Issue 1

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Ensemble pruning
    2. ensemble learning
    3. ensemble diversity
    4. deep learning

    Qualifiers

    • Research-article

    Funding Sources

    • NSF CISE
    • IBM Faculty Award, and a CISCO Edge AI

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)253
    • Downloads (Last 6 weeks)5
    Reflects downloads up to 25 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Effective Diversity Optimizations for High Accuracy Deep Ensembles2024 IEEE 6th International Conference on Cognitive Machine Intelligence (CogMI)10.1109/CogMI62246.2024.00044(278-287)Online publication date: 28-Oct-2024

    View Options

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media