Abstract
Broad Learning System (BLS), a type of neural network with a non-iterative training mechanism and adaptive network structure, has attracted much attention in recent years. In BLS, since the mapped features are obtained by mapping the training data based on a set of random weights, their quality is unstable, which in turn leads to the instability of the generalization ability of the model. To improve the diversity and stability of mapped features in BLS, we propose the BLS with Hybrid Features (BLSHF) algorithm in this study. Unlike original BLS, which uses a single uniform distribution to assign random values for the input weights of mapped feature nodes, BLSHF uses different distributions to initialize the mapped feature nodes in each group, thereby increasing the diversity of mapped features. This method enables BLSHF to extract high-level features from the original data better than the original BLS and further improves the feature extraction effect of the subsequent enhancement layer. Diverse features are beneficial to algorithms that use non-iterative training mechanisms, so BLSHF can achieve better generalization ability than BLS. We apply BLSHF to solve the problem of air quality evaluation, and the relevant experimental results empirically prove the effectiveness of this method. The learning mechanism of BLSHF can be easily applied to BLS and its variants to improve their generalization ability, which makes it have good application value.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Cao, W., Patwary, M.J., Yang, P., Wang, X., Ming, Z.: An initial study on the relationship between meta features of dataset and the initialization of nnrw. In: 2019 International Joint Conference on Neural Networks, IJCNN, pp. 1–8. IEEE (2019)
Cao, W., Wang, X., Ming, Z., Gao, J.: A review on neural networks with random weights. Neurocomputing 275, 278–287 (2018)
Chen, C.P., Liu, Z.: Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 29(1), 10–24 (2017)
Chen, C.P., Liu, Z., Feng, S.: Universal approximation capability of broad learning system and its structural variations. IEEE Trans. Neural Net. Learn. Syst. 30(4), 1191–1204 (2018)
Chu, F., Liang, T., Chen, C.P., Wang, X., Ma, X.: Weighted broad learning system and its application in nonlinear industrial process modeling. IEEE Trans. Neural Netw. Learn. Syst. 31(8), 3017–3031 (2019)
Feng, S., Chen, C.P.: Fuzzy broad learning system: a novel neuro-fuzzy model for regression and classification. IEEE Trans. Cybern. 50(2), 414–424 (2018)
Goodfellow, I., Bengio, Y., Courville, A.: Deep learning. MIT Press (2016)
Guo, P., Chen, C.P., Sun, Y.: An exact supervised learning for a three-layer supervised neural network. In: 1995 International Conference on Neural Information Processing, ICNIP, pp. 1041–1044 (1995)
Huang, G.-B., Zhu, Q.-Y., Siew, C.-K.: Extreme learning machine: a new learning scheme of feedforward neural networks. In: 2004 IEEE International Joint Conference on Neural Networks, IJCNN, vol. 2, pp. 985–990. IEEE (2004)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Commun. ACM 60(6), 84–90 (2017)
Liang, X., et al.: Assessing beijing’s pm2. 5 pollution: severity, weather impact, apec and winter heating. Proc. Royal Soc. A Math. Phys. Eng. Sci. 471(2182), 20150257 (2015)
Liu, Y., Cao, W., Ming, Z., Wang, Q., Zhang, J., Xu, Z.: Ensemble neural networks with random weights for classification problems. In: 2020 3rd International Conference on Algorithms, Computing and Artificial Intelligence, ACAI, pp. 1–5 (2020)
Shi, Z., Chen, X., Zhao, C., He, H., Stuphorn, V., Wu, D.: Multi-view broad learning system for primate oculomotor decision decoding. IEEE Trans. Neural Syst. Rehabil. Eng. 28(9), 1908–1920 (2020)
Wang, D., Li, M.: Stochastic configuration networks: fundamentals and algorithms. IEEE Trans. Cybern. 47(10), 3466–3479 (2017)
Xu, M., Han, M., Chen, C.P., Qiu, T.: Recurrent broad learning systems for time series prediction. IEEE Trans. Cybern. 50(4), 1405–1417 (2018)
Zhang, L., Suganthan, P.N.: A comprehensive evaluation of random vector functional link networks. Inf. Sci. 367, 1094–1105 (2016)
Zhang, L., Suganthan, P.N.: A survey of randomized algorithms for training neural networks. Inf. Sci. 364, 146–155 (2016)
Zhao, H., Zheng, J., Deng, W., Song, Y.: Semi-supervised broad learning system based on manifold regularization and broad network. IEEE Trans. Circuits Syst. I Regul. Pap. 67(3), 983–994 (2020)
Zhao, Z.-Q., Zheng, P., Xu, S.-T., Wu, X.: Object detection with deep learning: a review. IEEE Trans. Neural Netw. Learn. Syst. 30(11), 3212–3232 (2019)
Zou, W., Xia, Y., Cao, W.: Dense broad learning system based on conjugate gradient. In: 2020 International Joint Conference on Neural Networks, IJCNN, pp. 1–6. IEEE (2020)
Zou, W., Xia, Y., Cao, W., Ming, Z.: Broad learning system with proportional-integral-differential gradient descent. In: Qiu, M. (ed.) ICA3PP 2020. LNCS, vol. 12452, pp. 219–231. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-60245-1_15
Acknowledgment
This work was supported by National Natural Science Foundation of China (Grant No. 62106150), CAAC Key Laboratory of Civil Aviation Wide Surveillance and Safety Operation Management and Control Technology (Grant No. 202102), and CCF-NSFOCUS (Grant No. 2021001).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Cao, W., Li, D., Zhang, X., Qiu, M., Liu, Y. (2022). BLSHF: Broad Learning System with Hybrid Features. In: Memmi, G., Yang, B., Kong, L., Zhang, T., Qiu, M. (eds) Knowledge Science, Engineering and Management. KSEM 2022. Lecture Notes in Computer Science(), vol 13369. Springer, Cham. https://doi.org/10.1007/978-3-031-10986-7_53
Download citation
DOI: https://doi.org/10.1007/978-3-031-10986-7_53
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-10985-0
Online ISBN: 978-3-031-10986-7
eBook Packages: Computer ScienceComputer Science (R0)