Abstract
This article introduces a multiple classifier method to improve the performance of concatenate-designed neural networks, such as ResNet and DenseNet, with the purpose of alleviating the pressure on the final classifier. We give the design of the classifiers, which collects the features produced between the network sets, and present the constituent layers and the activation function for the classifiers, to calculate the classification score of each classifier. We use the \(L2 \left( \sqrt{e^x}\right)\) normalization method to obtain the classifier score instead of the \(Softmax\) normalization. We also determine the conditions that can enhance convergence. As a result, the proposed classifiers are able to improve the accuracy in the experimental cases significantly and show that the method not only has better performance than the original models, but also produces faster convergence. Moreover, our classifiers are general and can be applied to all classification related concatenate-designed network models.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
Generally, datasets CIFAR-10 and CIAR-100 can also reach complete convergence within 300 epochs, depending on the complexity of network architecture.
References
Bagheri MA, Gao Q, Escalera S (2013) A genetic-based subspace analysis method for improving error-correcting output coding. Pattern Recognit 46(10):2830–2839
Ball JE, Anderson DT, Chan CS (2017) A comprehensive survey of deep learning in remote sensing: Theories, tools and challenges for the community. arXiv: CoRR/abs/1709.00308
Bridle JS (1989) Probabilistic interpretation of feedforward classification network outputs, with relationships to statistical pattern recognition. In: Fogelman-Soulié F, Hérault J (eds) Proceedings of the NATO advanced research workshop on neurocomputing algorithms, architectures and applications, Les Arcs, France, February 27–March 3, 1989, Springer, NATO ASI Series, vol 68, pp 227–236, https://doi.org/10.1007/978-3-642-76153-9_28
Byun H, Lee S (2003) A survey on pattern recognition applications of support vector machines. Int J Pattern Recognit Artif Intell 17(3):459–486
Cao H, Bernard S, Heutte L, Sabourin R (2018) Dynamic voting in multi-view learning for radiomics applications. In: Bai X, Hancock ER, Ho TK, Wilson RC, Biggio B, Robles-Kelly A (eds) Structural, syntactic, and statistical pattern recognition–Joint IAPR international workshop, s+SSPR 2018, Beijing, China, August 17–19, 2018, Proceedings, Springer, Lecture Notes in Computer Science, vol 11004, pp 32–41. https://doi.org/10.1007/978-3-319-97785-0_4
Chen Y, Jin X, Kang B, Feng J, Yan S (2017) Sharing residual units through collective tensor factorization in deep neural networks. arXiv:CoRR/abs/1703.02180
Ciresan DC, Meier U, Gambardella LM, Schmidhuber J (2011) Convolutional neural network committees for handwritten character classification. In: 2011 International conference on document analysis and recognition, ICDAR 2011, Beijing, China, September 18–21, 2011, IEEE Computer Society, pp 1135–1139, https://doi.org/10.1109/ICDAR.2011.229
Collobert R, Sinz FH, Weston J, Bottou L (2006) Large scale transductive svms. J Mach Learn Res 7:1687–1712, http://jmlr.org/papers/v7/collobert06a.html
Dietterich TG (2000) Ensemble methods in machine learning. In: Kittler J, Roli F (eds) Multiple classifier systems, first international workshop, MCS 2000, Cagliari, Italy, June 21-23, 2000, Springer, Lecture Notes in Computer Science, vol 1857, pp 1–15. https://doi.org/10.1007/3-540-45014-9_1
Dietterich TG, Bakiri G (1995) Solving multiclass learning problems via error-correcting output codes. J Artif Intell Res 2:263–286
Díez-Pastor J, Rodríguez JJ, García-Osorio CI, Kuncheva LI (2015) Diversity techniques improve the performance of the best imbalance learning ensembles. Inf Sci 325:98–117
Elleuch M, Maalej R, Kherallah M (2016) A new design based-svm of the CNN classifier architecture with dropout for offline arabic handwritten recognition. In: Connolly M (ed) International conference on computational science 2016, ICCS 2016, 6-8 June 2016, San Diego, California, USA, Elsevier, Procedia Computer Science, vol 80, pp 1712–1723. https://doi.org/10.1016/j.procs.2016.05.512
Galar M, Fernández A, Tartas EB, Sola HB, Herrera F (2012) A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Trans Syst Man Cybern Part C 42(4):463–484
Gong S, Cristani M, Loy CC, Hospedales TM (2014) The re-identification challenge. In: Gong S, Cristani M, Yan S, Loy CC (eds) Person re-identification, advances in computer vision and pattern recognition, Springer, pp 1–20. https://doi.org/10.1007/978-1-4471-6296-4_1
Hammann M, Kraus M, Shafaei S, Knoll AC (2020) Identity recognition in intelligent cars with behavioral data and lstm-resnet classifier. arXiv:CoRR/abs/2003.00770
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: 2016 IEEE conference on computer vision and pattern recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, IEEE Computer Society, pp 770–778. https://doi.org/10.1109/CVPR.2016.90
Hershey S, Chaudhuri S, Ellis DPW, Gemmeke JF, Jansen A, Moore RC, Plakal M, Platt D, Saurous RA, Seybold B, Slaney M, Weiss RJ, Wilson KW (2017) CNN architectures for large-scale audio classification. In: 2017 IEEE international conference on acoustics, speech and signal processing, ICASSP 2017, New Orleans, LA, USA, March 5-9, 2017, IEEE, pp 131–135. https://doi.org/10.1109/ICASSP.2017.7952132
Huang G, Liu Z, van der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: 2017 IEEE conference on computer vision and pattern recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, IEEE Computer Society, pp 2261–2269. https://doi.org/10.1109/CVPR.2017.243
Ioffe S, Szegedy C (2015) Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: Bach FR, Blei DM (eds) Proceedings of the 32nd international conference on machine learning, ICML 2015, Lille, France, 6-11 July 2015, JMLR.org, JMLR Workshop and Conference Proceedings, vol 37, pp 448–456, http://proceedings.mlr.press/v37/ioffe15.html
Jiang J, Trundle PR, Ren J (2010) Medical image analysis with artificial neural networks. Comput Medical Imaging Graph 34(8):617–631
Kim DK, Chen T (2015) Deep neural network for real-time autonomous indoor navigation. arXiv:CoRR/abs/1511.04668
Kingma DP, Ba J (2015) Adam: A method for stochastic optimization. In: Bengio Y, LeCun Y (eds) 3rd international conference on learning representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, conference track proceedings, arxiv:abs/1412.6980
Krizhevsky A (2009) Learning multiple layers of features from tiny images. Tech. rep
Krizhevsky A, Sutskever I, Hinton GE (2017) Imagenet classification with deep convolutional neural networks. Commun ACM 60(6):84–90
Lai K, Liu D, Chang S, Chen M (2015) Learning sample specific weights for late fusion. IEEE Trans Image Process 24(9):2772–2783
Lawrence S, Giles CL, Tsoi AC, Back AD (1997) Face recognition: a convolutional neural-network approach. IEEE Trans Neural Netw 8(1):98–113
Lecun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324. https://doi.org/10.1109/5.726791
Liu X, Gao J, He X, Deng L, Duh K, Wang Y (2015) Representation learning using multi-task deep neural networks for semantic classification and information retrieval. In: Mihalcea R, Chai JY, Sarkar A (eds) NAACL HLT 2015, The 2015 conference of the north american chapter of the association for computational linguistics: human language technologies, Denver, Colorado, USA, May 31 - June 5, 2015, The Association for Computational Linguistics, pp 912–921. https://doi.org/10.3115/v1/n15-1092
Mandelbaum A, Weinshall D (2017) Distance-based confidence score for neural network classifiers. arXiv:CoRR/abs/1709.09844
Naranjo L, Perez CJ, Martín J, Campos-Roca Y (2017) A two-stage variable selection and classification approach for Parkinsons disease detection by using voice recording replications. Comput Methods Programs Biomed 142:147–156
Nawaz W, Ahmed S, Tahir A, Khan HA (2018) Classification of breast cancer histology images using ALEXNET. In: Campilho A, Karray F, ter Haar Romeny BM (eds) Image analysis and recognition - 15th international conference, ICIAR 2018, Póvoa de Varzim, Portugal, June 27-29, 2018, Proceedings, Springer, Lecture Notes in Computer Science, vol 10882, pp 869–876. https://doi.org/10.1007/978-3-319-93000-8_99
Nortje L, Kamper H (2020) Unsupervised vs. transfer learning for multimodal one-shot matching of speech and images. In: Meng H, Xu B, Zheng TF (eds) Interspeech 2020, 21st annual conference of the international speech communication association, virtual event, Shanghai, China, 25-29 October 2020, ISCA, pp 2712–2716. https://doi.org/10.21437/Interspeech.2020-0087
Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, Killeen T, Lin Z, Gimelshein N, Antiga L, Desmaison A, Köpf A, Yang E, DeVito Z, Raison M, Tejani A, Chilamkurthy S, Steiner B, Fang L, Bai J, Chintala S (2019) Pytorch: An imperative style, high-performance deep learning library. In: Wallach HM, Larochelle H, Beygelzimer A, d’Alché-Buc F, Fox EB, Garnett R (eds) Advances in neural information processing systems 32: annual conference on neural information processing systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp 8024–8035, https://proceedings.neurips.cc/paper/2019/hash/bdbca288fee7f92f2bfa9f7012727740-Abstract.html
Qian Y, Ye M, Zhou J (2013) Hyperspectral image classification based on structured sparse logistic regression and three-dimensional wavelet texture features. IEEE Trans Geosci Remote Sens 51(4–2):2276–2291
Ranjan R, Sankaranarayanan S, Castillo CD, Chellappa R (2017) An all-in-one convolutional neural network for face analysis. In: 12th IEEE international conference on automatic face & gesture recognition, FG 2017, Washington, DC, USA, May 30–June 3, 2017, IEEE Computer Society, pp 17–24. https://doi.org/10.1109/FG.2017.137
Ren Z, Lee YJ (2018) Cross-domain self-supervised multi-task feature learning using synthetic imagery. In: 2018 IEEE conference on computer vision and pattern recognition, CVPR 2018, Salt Lake City, UT, USA, June 18–22, 2018, IEEE Computer Society, pp 762–771, DOIurl10.1109/CVPR.2018.00086, http://openaccess.thecvf.com/content_cvpr_2018/html/Ren_Cross-Domain_Self-Supervised_Multi-Task_CVPR_2018_paper.html
Sandler M, Howard AG, Zhu M, Zhmoginov A, Chen L (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: 2018 IEEE conference on computer vision and pattern recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, IEEE Computer Society, pp 4510–4520. DOIurl10.1109/CVPR.2018.00474, http://openaccess.thecvf.com/content_cvpr_2018/html/Sandler_MobileNetV2_Inverted_Residuals_CVPR_2018_paper.html
Shorten C, Khoshgoftaar TM (2019) A survey on image data augmentation for deep learning. J Big Data 6:60
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: Bengio Y, LeCun Y (eds) 3rd international conference on learning representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, conference track proceedings, arxiv:abs/1409.1556
Sun Y, Liang D, Wang X, Tang X (2015) Deepid3: Face recognition with very deep neural networks. arXiv:CoRR/abs/1502.00873
Szegedy C, Liu W, Jia Y, Sermanet P, Reed SE, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: IEEE conference on computer vision and pattern recognition, CVPR 2015, Boston, MA, USA, June 7-12, 2015, IEEE Computer Society, pp 1–9. DOIurl10.1109/CVPR.2015.7298594
Takemura A, Shimizu A, Hamamoto K (2010) Discrimination of breast tumors in ultrasonic images using an ensemble classifier based on the adaboost algorithm with feature selection. IEEE Trans Med Imaging 29(3):598–609
Tan M, Le QV (2019) Efficientnet: Rethinking model scaling for convolutional neural networks. In: Chaudhuri K, Salakhutdinov R (eds) Proceedings of the 36th international conference on machine learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, PMLR, proceedings of machine learning research, vol 97, pp 6105–6114, http://proceedings.mlr.press/v97/tan19a.html
Tian Y, Pei K, Jana S, Ray B (2018) Deeptest: automated testing of deep-neural-network-driven autonomous cars. In: Chaudron M, Crnkovic I, Chechik M, Harman M (eds) Proceedings of the 40th international conference on software engineering, ICSE 2018, Gothenburg, Sweden, May 27–June 03, 2018, ACM, pp 303–314, https://doi.org/10.1145/3180155.3180220
Ueda N (2000) Optimal linear combination of neural networks for improving classification performance. IEEE Trans Pattern Anal Mach Intell 22(2):207–215
Vapnik V (1999) An overview of statistical learning theory. IEEE Trans Neural Netw 10(5):988–999
Wang L, Lee C, Tu Z, Lazebnik S (2015) Training deeper convolutional networks with deep supervision. arXiv:CoRR/abs/1505.02496
Wei Y, Zhang Y, Huang J, Yang Q (2018) Transfer learning via learning to transfer. In: Dy JG, Krause A (eds) Proceedings of the 35th international conference on machine learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10–15, 2018, PMLR, proceedings of machine learning research, vol 80, pp 5072–5081, http://proceedings.mlr.press/v80/wei18a.html
Wu Z, Shen C, van den Hengel A (2019) Wider or deeper: revisiting the resnet model for visual recognition. Pattern Recognit 90:119–133
Xiao L, Yan Q, Deng S (2017) Scene classification with improved alexnet model. In: Li T, Martínez-López L, Li Y (eds) 12th international conference on intelligent systems and knowledge engineering, ISKE 2017, Nanjing, China, November 24–26, 2017, IEEE, pp 1–6. https://doi.org/10.1109/ISKE.2017.8258820
Xie S, Girshick RB, Dollár P, Tu Z, He K (2017) Aggregated residual transformations for deep neural networks. In: 2017 IEEE conference on computer vision and pattern recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, IEEE Computer Society, pp 5987–5995. https://doi.org/10.1109/CVPR.2017.634
Xu B, Guo X, Ye Y, Cheng J (2012) An improved random forest classifier for text categorization. J Comput 7(12):2913–2920
Yu F, Wang D, Shelhamer E, Darrell T (2018) Deep layer aggregation. In: 2018 IEEE conference on computer vision and pattern recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, IEEE Computer Society, pp 2403–2412, https://doi.org/10.1109/CVPR.2018.00255, http://openaccess.thecvf.com/content_cvpr_2018/html/Yu_Deep_Layer_Aggregation_CVPR_2018_paper.html
Zhang Y, Górriz JM, Dong Z (2021) Deep learning in medical image analysis. J Imaging 7(4):74
Zheng X, Lin L, Liu B, Xiao Y, Xiong X (2020) A multi-task transfer learning method with dictionary learning. Knowl Based Syst 191:105233
Zhong Z, Zheng L, Kang G, Li S, Yang Y (2020) Random erasing data augmentation. In: The Thirty-Fourth AAAI conference on artificial intelligence, AAAI 2020, the thirty-second innovative applications of artificial intelligence conference, IAAI 2020, The Tenth AAAi symposium on educational advances in artificial intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, AAAI Press, pp 13001–13008, https://aaai.org/ojs/index.php/AAAI/article/view/7000
Funding
This work was funded by The Science and Technology Development Fund, Macao SAR (File no. 0001/2018/AFJ).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors have no conflicts of interest to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Chan, KH., Im, SK. & Ke, W. Multiple classifier for concatenate-designed neural network. Neural Comput & Applic 34, 1359–1372 (2022). https://doi.org/10.1007/s00521-021-06462-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-021-06462-0