A Review on Large-Scale Data Processing with Parallel and Distributed Randomized Extreme Learning Machine Neural Networks
Abstract
:1. Introduction
- An overview of the basic structure of ELM and the variants with parallel and distributed computing to solve problems with large-scale data sets.
- A discussion of advances in solving SLE using parallel and distributed computing to address high-dimensional arrays.
- A description of the parallel and distributed tools used to improve the performance of the ELM algorithm and its variants by solving problems associated with training time and memory capacity.
- A summary of the evolution in the last decade of the ELM algorithm and its variants combined with parallel and distributed tools.
2. Related Reviews about Extreme Learning Machine
3. Background
3.1. Moore–Penrose Generalized Inverse
3.2. Standard Model of Extreme Learning Machine
3.3. Randomized Feedforward Neural Networks and ELM’s Origin
3.4. ELM Variants Implemented by Using Distributed and Parallel Computing
3.5. Other Variants
4. Methods for Solving Linear Systems with Parallel and Distributed Computing
5. Review of Distributed and Parallel Systems for Extreme Learning Machine
5.1. MapReduce
5.2. Spark
5.3. Graphics Processing Unit
5.4. Other Tools and Technologies for Distributed and Parallel Computing
6. Discussion
7. Conclusions
- Researchers who wish to use distributed and parallel computing in ELM, its variants, and other randomized feedforward neural network models can use this work as a reference to identify the technologies and tools that have been implemented so far.
- According to the large amount of time involved in computing the MPGI matrix, we have presented an updated review of the most widely used methods and their implementation with distributed and parallel architectures and tools.
- The review on distributed and parallel methods for computing the MPGI matrix is relevant to accelerate the training of ELM, its variants, and other randomized feedforward neural network models.
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Schmidt, W.F.; Kraaijveld, M.A.; Duin, R.P. Feed forward neural networks with random weights. In Proceedings of the 11th IAPR International Conference on Pattern Recognition. Vol. II. Conference B: Pattern Recognition Methodology and Systems, The Hague, The Netherlands, 30 August–3 September 1992; IEEE: Piscataway, NJ, USA, 1992; pp. 1–4. [Google Scholar] [CrossRef]
- Pao, Y.H.; Takefuji, Y. Functional-link net computing: Theory, system architecture, and functionalities. Computer 1992, 25, 76–79. [Google Scholar] [CrossRef]
- Pao, Y.H.; Park, G.H.; Sobajic, D.J. Learning and generalization characteristics of the random vector functional-link net. Neurocomputing 1994, 6, 163–180. [Google Scholar] [CrossRef]
- Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No. 04CH37541), Budapest, Hungary, 25–29 July 2004; IEEE: Piscataway, NJ, USA, 2004; pp. 985–990. [Google Scholar] [CrossRef]
- Huang, G.B.; Wang, D.H.; Lan, Y. Extreme learning machines: A survey. Int. J. Mach. Learn. Cybern. 2011, 2, 107–122. [Google Scholar] [CrossRef]
- Ahmadi, M.; Soofiabadi, M.; Nikpour, M.; Naderi, H.; Abdullah, L.; Arandian, B. Developing a deep neural network with fuzzy wavelets and integrating an inline PSO to predict energy consumption patterns in urban buildings. Mathematics 2022, 10, 1270. [Google Scholar] [CrossRef]
- Sharifi, A.; Ahmadi, M.; Mehni, M.A.; Ghoushchi, S.J.; Pourasad, Y. Experimental and numerical diagnosis of fatigue foot using convolutional neural network. Comput. Methods Biomech. Biomed. Eng. 2021, 24, 1828–1840. [Google Scholar] [CrossRef]
- Ahmadi, M.; Ahangar, F.D.; Astaraki, N.; Abbasi, M.; Babaei, B. FWNNet: Presentation of a new classifier of brain tumor diagnosis based on fuzzy logic and the wavelet-based neural network using machine-learning methods. Comput. Intell. Neurosci. 2021, 2021, 8542637. [Google Scholar] [CrossRef] [PubMed]
- Nomani, A.; Ansari, Y.; Nasirpour, M.H.; Masoumian, A.; Pour, E.S.; Valizadeh, A. PSOWNNs-CNN: A Computational Radiology for Breast Cancer Diagnosis Improvement Based on Image Processing Using Machine Learning Methods. Comput. Intell. Neurosci. 2022, 2022, 5667264. [Google Scholar] [CrossRef]
- Zangeneh Soroush, M.; Tahvilian, P.; Nasirpour, M.H.; Maghooli, K.; Sadeghniiat-Haghighi, K.; Vahid Harandi, S.; Abdollahi, Z.; Ghazizadeh, A.; Jafarnia Dabanloo, N. EEG artifact removal using sub-space decomposition, nonlinear dynamics, stationary wavelet transform and machine learning algorithms. Front. Physiol. 2022, 13, 1572. [Google Scholar] [CrossRef]
- Huérfano-Maldonado, Y.; Mora, M.; Vilches, K.; Hernández-García, R.; Gutiérrez, R.; Vera, M. A comprehensive review of extreme learning machine on medical imaging. Neurocomputing 2023, 556, 126618. [Google Scholar] [CrossRef]
- Patil, H.; Sharma, K. Extreme learning machine: A comprehensive survey of theories & algorithms. In Proceedings of the 2023 International Conference on Computational Intelligence and Sustainable Engineering Solutions (CISES), Greater Noida, India, 28–30 April 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 749–756. [Google Scholar] [CrossRef]
- Kaur, R.; Roul, R.K.; Batra, S. Multilayer extreme learning machine: A systematic review. Multimed. Tools Appl. 2023, 82, 40269–40307. [Google Scholar] [CrossRef]
- Vásquez-Coronel, J.A.; Mora, M.; Vilches, K. A Review of multilayer extreme learning machine neural networks. Artif. Intell. Rev. 2023, 56, 13691–13742. [Google Scholar] [CrossRef]
- Wang, J.; Lu, S.; Wang, S.H.; Zhang, Y.D. A review on extreme learning machine. Multimed. Tools Appl. 2022, 81, 41611–41660. [Google Scholar] [CrossRef]
- Zheng, X.; Li, P.; Wu, X. Data Stream Classification Based on Extreme Learning Machine: A Review. Big Data Res. 2022, 30, 100356. [Google Scholar] [CrossRef]
- Martínez, D.; Zabala-Blanco, D.; Ahumada-García, R.; Azurdia-Meza, C.A.; Flores-Calero, M.; Palacios-Jativa, P. Review of extreme learning machines for the identification and classification of fingerprint databases. In Proceedings of the 2022 IEEE Colombian Conference on Communications and Computing (COLCOM), Cali, Colombia, 27–29 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar] [CrossRef]
- Kaur, M.; Das, D.; Mishra, S.P. Survey and evaluation of extreme learning machine on TF-IDF feature for sentiment analysis. In Proceedings of the 2022 International Conference on Machine Learning, Computer Systems and Security (MLCSS), Bhubaneswar, India, 5–6 August 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 247–252. [Google Scholar] [CrossRef]
- Nilesh, R.; Sunil, W. Review of Optimization in Improving Extreme Learning Machine. EAI Endorsed Trans. Ind. Netw. Intell. Syst. 2021, 8, e2. [Google Scholar] [CrossRef]
- Mujal, P.; Martínez-Peña, R.; Nokkala, J.; García-Beni, J.; Giorgi, G.L.; Soriano, M.C.; Zambrini, R. Opportunities in quantum reservoir computing and extreme learning machines. Adv. Quantum Technol. 2021, 4, 2100027. [Google Scholar] [CrossRef]
- Nilesh, R.; Sunil, W. Improving extreme learning machine through optimization a review. In Proceedings of the 2021 7th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 19–20 March 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 906–912. [Google Scholar] [CrossRef]
- Rodrigues, I.R.; da Silva Neto, S.R.; Kelner, J.; Sadok, D.; Endo, P.T. Convolutional Extreme Learning Machines: A Systematic Review. Informatics 2021, 8, 33. [Google Scholar] [CrossRef]
- Saldaña-Olivas, E.; Huamán-Tuesta, J.R. Extreme learning machine for business sales forecasts: A systematic review. In Smart Innovation, Systems and Technologies, Proceedings of the 5th Brazilian Technology Symposium (BTSym 2019), Campinas, Brazil, 22–24 October 2019; Iano, Y., Arthur, R., Saotome, O., Kemper, G., Padilha França, R., Eds.; Springer: Sao Paulo, Barzil, 2021; pp. 87–96. [Google Scholar] [CrossRef]
- Wang, Z.; Luo, Y.; Xin, J.; Zhang, H.; Qu, L.; Wang, Z.; Yao, Y.; Zhu, W.; Wang, X. Computer-Aided Diagnosis Based on Extreme Learning Machine: A Review. IEEE Access 2020, 8, 141657–141673. [Google Scholar] [CrossRef]
- Wang, Z.; Sui, L.; Xin, J.; Qu, L.; Yao, Y. A Survey of Distributed and Parallel Extreme Learning Machine for Big Data. IEEE Access 2020, 8, 201247–201258. [Google Scholar] [CrossRef]
- Alaba, P.A.; Popoola, S.I.; Olatomiwa, L.; Akanle, M.B.; Ohunakin, O.S.; Adetiba, E.; Alex, O.D.; Atayero, A.A.; Daud, W.M.A.W. Towards a more efficient and cost-sensitive extreme learning machine: A state-of-the-art review of recent trend. Neurocomputing 2019, 350, 70–90. [Google Scholar] [CrossRef]
- Yibo, L.; Fang, L.; Qi, C. A Review of the Research on the Prediction Model of Extreme Learning Machine. J. Phys. Conf. Ser. 2019, 1213, 042013. [Google Scholar] [CrossRef]
- Li, L.; Sun, R.; Cai, S.; Zhao, K.; Zhang, Q. A review of improved extreme learning machine methods for data stream classification. Multimed. Tools Appl. 2019, 78, 33375–33400. [Google Scholar] [CrossRef]
- Eshtay, M.; Faris, H.; Obeid, N. Metaheuristic-based extreme learning machines: A review of design formulations and applications. Int. J. Mach. Learn. Cybern. 2019, 10, 1543–1561. [Google Scholar] [CrossRef]
- Ghosh, S.; Mukherjee, H.; Obaidullah, S.M.; Santosh, K.; Das, N.; Roy, K. A survey on extreme learning machine and evolution of its variants. In Proceedings of the Recent Trends in Image Processing and Pattern Recognition. Second International Conference, RTIP2R 2018, Solapur, India, 21–22 December 2018; Santosh, K.C., Hegadi, R.S., Eds.; Springer: Singapore, 2019; Volume 1035, pp. 572–583. [Google Scholar] [CrossRef]
- Zhang, S.; Tan, W.; Li, Y. A survey of online sequential extreme learning machine. In Proceedings of the 2018 5th International Conference on Control, Decision and Information Technologies (CoDIT), Thessaloniki, Greece, 10–13 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 45–50. [Google Scholar] [CrossRef]
- Alade, O.A.; Selamat, A.; Sallehuddin, R. A review of advances in extreme learning machine techniques and its applications. In Proceedings of the Recent Trends in Information and Communication Technology, Johor Bahru, Malaysia, 23–24 April 2017; Saeed, F., Gazem, N., Patnaik, S., Saed Balaid, A.S., Mohammed, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2017; pp. 885–895. [Google Scholar] [CrossRef]
- Salaken, S.M.; Khosravi, A.; Nguyen, T.; Nahavandi, S. Extreme learning machine based transfer learning algorithms: A survey. Neurocomputing 2017, 267, 516–524. [Google Scholar] [CrossRef]
- Albadra, M.A.A.; Tiun, S. Extreme learning machine: A review. Int. J. Appl. Eng. Res. 2017, 12, 4610–4623. [Google Scholar]
- Ali, M.H.; Zolkipli, M.F. Review on hybrid extreme learning machine and genetic algorithm to work as intrusion detection system in cloud computing. ARPN J. Eng. Appl. Sci. 2016, 11, 460–464. [Google Scholar]
- Huang, G.; Huang, G.B.; Song, S.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef]
- Cao, J.; Lin, Z. Extreme Learning Machines on High Dimensional and Large Data Applications: A Survey. Math. Probl. Eng. 2015, 2015, 103796. [Google Scholar] [CrossRef]
- Ding, S.; Zhao, H.; Zhang, Y.; Xu, X.; Nie, R. Extreme learning machine: Algorithm, theory and applications. Artif. Intell. Rev. 2015, 44, 103–115. [Google Scholar] [CrossRef]
- Deng, C.; Huang, G.; Xu, J.; Tang, J. Extreme learning machines: New trends and applications. Sci. China Inf. Sci. 2015, 58, 1–16. [Google Scholar] [CrossRef]
- Ding, S.; Xu, X.; Nie, R. Extreme learning machine and its applications. Neural Comput. Appl. 2014, 25, 549–556. [Google Scholar] [CrossRef]
- Liang, N.Y.; Huang, G.B.; Saratchandran, P.; Sundararajan, N. A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans. Neural Netw. 2006, 17, 1411–1423. [Google Scholar] [CrossRef]
- Ali, M.H.; Fadlizolkipi, M.; Firdaus, A.; Khidzir, N.Z. A hybrid particle swarm optimization-extreme learning machine approach for intrusion detection system. In Proceedings of the 2018 IEEE Student Conference on Research and Development (SCOReD), Selangor, Malaysia, 26–28 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–4. [Google Scholar] [CrossRef]
- Lyche, T. Numerical Linear Algebra and Matrix Factorizations; Springer: Oslo, Norway, 2020; Volume 22. [Google Scholar]
- Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
- Zhang, L.; Suganthan, P.N. A survey of randomized algorithms for training neural networks. Inf. Sci. 2016, 364–365, 146–155. [Google Scholar] [CrossRef]
- Suganthan, P.N.; Katuwal, R. On the origins of randomization-based feedforward neural networks. Appl. Soft Comput. 2021, 105, 107239. [Google Scholar] [CrossRef]
- Malik, A.K.; Gao, R.; Ganaie, M.; Tanveer, M.; Suganthan, P.N. Random vector functional link network: Recent developments, applications, and future directions. Appl. Soft Comput. 2023, 143, 110377. [Google Scholar] [CrossRef]
- Safaei, A.; Wu, Q.J.; Akilan, T.; Yang, Y. System-on-a-Chip (SoC)-Based Hardware Acceleration for an Online Sequential Extreme Learning Machine (OS-ELM). IEEE Trans.-Comput.-Aided Des. Integr. Circuits Syst. 2019, 38, 2127–2138. [Google Scholar] [CrossRef]
- Grim, L.F.L.; Barajas, J.A.B.; Gradvohl, A.L.S. Implementações paralelas para o algoritmo Online Sequential Extreme Learning Machine aplicado à previsão de material particulado. Rev. Bras. Comput. Apl. 2019, 11, 13–21. [Google Scholar] [CrossRef]
- Zehai, G.; Cunbao, M.; Jianfeng, Z.; Weijun, X. Remaining useful life prediction of integrated modular avionics using ensemble enhanced online sequential parallel extreme learning machine. Int. J. Mach. Learn. Cybern. 2021, 12, 1893–1911. [Google Scholar] [CrossRef]
- Polat, Ö.; Kayhan, S.K. GPU-accelerated and mixed norm regularized online extreme learning machine. Concurr. Comput. Pract. Exp. 2022, 34, e6967. [Google Scholar] [CrossRef]
- Vovk, V. Kernel ridge regression. In Empirical Inference; Schölkopf, B., Luo, Z., Vovk, V., Eds.; Springer: Berlin, Germany, 2013; pp. 105–116. [Google Scholar] [CrossRef]
- Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme Learning Machine for Regression and Multiclass Classification. IEEE Trans. Syst. Man, Cybern. Part Cybern. 2011, 42, 513–529. [Google Scholar] [CrossRef]
- Deng, W.Y.; Ong, Y.S.; Tan, P.S.; Zheng, Q.H. Online sequential reduced kernel extreme learning machine. Neurocomputing 2016, 174, 72–84. [Google Scholar] [CrossRef]
- Wu, L.; Peng, Y.; Fan, J.; Wang, Y.; Huang, G. A novel kernel extreme learning machine model coupled with K-means clustering and firefly algorithm for estimating monthly reference evapotranspiration in parallel computation. Agric. Water Manag. 2021, 245, 106624. [Google Scholar] [CrossRef]
- Huang, G.B.; Chen, L.; Siew, C.K. Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Netw. 2006, 17, 879–892. [Google Scholar] [CrossRef] [PubMed]
- Rong, H.J.; Ong, Y.S.; Tan, A.H.; Zhu, Z. A fast pruned-extreme learning machine for classification problem. Neurocomputing 2008, 72, 359–366. [Google Scholar] [CrossRef]
- Zhu, Q.Y.; Qin, A.K.; Suganthan, P.N.; Huang, G.B. Evolutionary extreme learning machine. Pattern Recognit. 2005, 38, 1759–1763. [Google Scholar] [CrossRef]
- Gelvez-Almeida, E.; Baldera-Moreno, Y.; Huérfano, Y.; Vera, M.; Mora, M.; Barrientos, R. Parallel methods for linear systems solution in extreme learning machines: An overview. J. Phys. Conf. Ser. 2020, 1702, 012017. [Google Scholar] [CrossRef]
- Lu, S.; Wang, X.; Zhang, G.; Zhou, X. Effective algorithms of the Moore-Penrose inverse matrices for extreme learning machine. Intell. Data Anal. 2015, 19, 743–760. [Google Scholar] [CrossRef]
- Young, D.M. Iterative Solution of Large Linear Systems; Elsevier: Orlando, FL, USA, 2014. [Google Scholar]
- Li, J.; Li, L.; Wang, Q.; Xue, W.; Liang, J.; Shi, J. Parallel optimization and application of unstructured sparse triangular solver on new generation of sunway architecture. Parallel Comput. 2024, 120, 103080. [Google Scholar] [CrossRef]
- Gelvez-Almeida, E.; Barrientos, R.J.; Vilches-Ponce, K.; Mora, M. A Parallel Computing Method for the Computation of the Moore–Penrose Generalized Inverse for Shared-Memory Architectures. IEEE Access 2023, 11, 134834–134845. [Google Scholar] [CrossRef]
- Lukyanenko, D. Parallel algorithm for solving overdetermined systems of linear equations, taking into account round-off errors. Algorithms 2023, 16, 242. [Google Scholar] [CrossRef]
- Suzuki, K.; Fukaya, T.; Iwashita, T. A novel ILU preconditioning method with a block structure suitable for SIMD vectorization. J. Comput. Appl. Math. 2023, 419, 114687. [Google Scholar] [CrossRef]
- Sabelfeld, K.K.; Kireev, S.; Kireeva, A. Parallel implementations of randomized vector algorithm for solving large systems of linear equations. J. Supercomput. 2023, 79, 10555–10569. [Google Scholar] [CrossRef]
- Catalán, S.; Herrero, J.R.; Igual, F.D.; Quintana-Ortí, E.S.; Rodríguez-Sánchez, R. Fine-grain task-parallel algorithms for matrix factorizations and inversion on many-threaded CPUs. Concurr. Comput. Pract. Exp. 2022, 35, e6999. [Google Scholar] [CrossRef]
- Rivera-Zamarripa, L.; Adj, G.; Cruz-Cortés, N.; Aguilar-Ibañez, C.; Rodríguez-Henríquez, F. A Parallel Strategy for Solving Sparse Linear Systems Over Finite Fields. Comput. Sist. 2022, 26, 493–504. [Google Scholar] [CrossRef]
- Li, K.; Han, X. A distributed Gauss-Newton method for distribution system state estimation. Int. J. Electr. Power Energy Syst. 2022, 136, 107694. [Google Scholar] [CrossRef]
- Hwang, H.S.; Ro, J.H.; Park, C.Y.; You, Y.H.; Song, H.K. Efficient Gauss-Seidel Precoding with Parallel Calculation in Massive MIMO Systems. CMC-Comput. Mater. Contin. 2022, 70, 491–504. [Google Scholar] [CrossRef]
- Catalán, S.; Igual, F.D.; Rodríguez-Sánchez, R.; Herrero, J.R.; Quintana-Ortí, E.S. A New Generation of Task-Parallel Algorithms for Matrix Inversion in Many-Threaded CPUs. In Proceedings of the 12th International Workshop on Programming Models and Applications for Multicores and Manycores, Association for Computing Machinery, Virtual, 22 February 2021; pp. 1–10. [Google Scholar] [CrossRef]
- Marrakchi, S.; Jemni, M. Parallel gaussian elimination of symmetric positive definite band matrices for shared-memory multicore architectures. RAIRO Oper. Res. 2021, 55, 905–927. [Google Scholar] [CrossRef]
- Lu, Y.; Luo, Y.; Lian, H.; Jin, Z.; Liu, W. Implementing LU and Cholesky factorizations on artificial intelligence accelerators. CCF Trans. High Perform. Comput. 2021, 3, 286–297. [Google Scholar] [CrossRef]
- Lee, W.K.; Achar, R. GPU-Accelerated Adaptive PCBSO Mode-Based Hybrid RLA for Sparse LU Factorization in Circuit Simulation. IEEE Trans.-Comput.-Aided Des. Integr. Circuits Syst. 2021, 40, 2320–2330. [Google Scholar] [CrossRef]
- Zhang, X.W.; Zuo, L.; Li, M.; Guo, J.X. High-throughput FPGA implementation of matrix inversion for control systems. IEEE Trans. Ind. Electron. 2021, 68, 6205–6216. [Google Scholar] [CrossRef]
- Rubensson, E.H.; Artemov, A.G.; Kruchinina, A.; Rudberg, E. Localized inverse factorization. IMA J. Numer. Anal. 2021, 41, 729–763. [Google Scholar] [CrossRef]
- Rodriguez Borbon, J.M.; Huang, J.; Wong, B.M.; Najjar, W. Acceleration of Parallel-Blocked QR Decomposition of Tall-and-Skinny Matrices on FPGAs. ACM Trans. Archit. Code Optim. TACO 2021, 18, 27. [Google Scholar] [CrossRef]
- Duan, T.; Dinavahi, V. A novel linking-domain extraction decomposition method for parallel electromagnetic transient simulation of large-scale AC/DC networks. IEEE Trans. Power Deliv. 2021, 36, 957–965. [Google Scholar] [CrossRef]
- Shäfer, F.; Katzfuss, M.; Owhadi, H. Sparse Cholesky Factorization by Kullback-Leibler Minimization. SIAM J. Sci. Comput. 2021, 43, A2019–A2046. [Google Scholar] [CrossRef]
- Boffi, D.; Lu, Z.; Pavarino, L.F. Iterative ILU preconditioners for linear systems and eigenproblems. J. Comput. Math. 2021, 39, 633–654. [Google Scholar] [CrossRef]
- Ahmadi, A.; Manganiello, F.; Khademi, A.; Smith, M.C. A Parallel Jacobi-Embedded Gauss-Seidel Method. IEEE Trans. Parallel Distrib. Syst. 2021, 32, 1452–1464. [Google Scholar] [CrossRef]
- Liu, Y.; Sid-Lakhdar, W.; Rebrova, E.; Ghysels, P.; Li, X.S. A parallel hierarchical blocked adaptive cross approximation algorithm. Int. J. High Perform. Comput. Appl. 2020, 34, 394–408. [Google Scholar] [CrossRef]
- Davis, T.A.; Duff, I.S.; Nakov, S. Design and implementation of a parallel markowitz threshold algorithm. SIAM J. Matrix Anal. Appl. 2020, 41, 573–590. [Google Scholar] [CrossRef]
- Yang, X.; Wang, N.; Xu, L. A parallel Gauss-Seidel method for convex problems with separable structure. Numer. Algebr. Control. Optim. 2020, 10, 557–570. [Google Scholar] [CrossRef]
- Li, R.; Zhang, C. Efficient parallel implementations of sparse triangular solves for GPU architectures. In Proceedings of the 2020 SIAM Conference on Parallel Processing for Scientific Computing, SIAM, Washington, DC, USA, 12–15 February 2020; pp. 106–117. [Google Scholar] [CrossRef]
- Singh, N.; Ma, L.; Yang, H.; Solomonik, E. Comparison of Accuracy and Scalability of Gauss-Newton and Alternating Least Squares for CP Decomposition. arXiv 2020, arXiv:1910.12331. [Google Scholar] [CrossRef]
- Alyahya, H.; Mehmood, R.; Katib, I. Parallel iterative solution of large sparse linear equation systems on the intel MIC architecture. In Smart Infrastructure and Applications; Mehmood, R., See, S., Katib, I., Chlamtac, I., Eds.; Springer: Cham, Switzerland, 2020; pp. 377–407. [Google Scholar] [CrossRef]
- Huang, G.H.; Xu, Y.Z.; Yi, X.W.; Xia, M.; Jiao, Y.Y.; Zhang, S. Highly efficient iterative methods for solving linear equations of three-dimensional sphere discontinuous deformation analysis. Int. J. Numer. Anal. Methods Geomech. 2020, 44, 1301–1314. [Google Scholar] [CrossRef]
- Kirk, D.B.; Mei, W. Hwu, W. Programming Massively Parallel Processors: A Hands-On Approach, 3rd ed.; Morgan Kaufmann: Cambridge, UK, 2016. [Google Scholar]
- Chapman, B.; Jost, G.; Pas, R.V.D. Using OpenMP: Portable Shared Memory Parallel Programming; The MIT Press: London, UK, 2008. [Google Scholar]
- Xianyi, Z.; Kroeker, M. OpenBLAS: An Optimized BLAS Library. 2022. Available online: https://www.openblas.net (accessed on 20 September 2022).
- University of Tennessee; University of California; University of Colorado Denver; NAG Ltd. LAPACK—Linear Algebra PACKage. Netlib Repository at UTK and ORNL. 2022. Available online: http://www.netlib.org/lapack/ (accessed on 15 September 2022).
- Gropp, W.; Lusk, E.; Skjellum, A. Using MPI: Portable Parallel Programming with the Message-Passing Interface (Scientific and Engineering Computation Series), 3rd ed.; The MIT Press: London, UK, 2014. [Google Scholar]
- Intel Corporation. Intel oneAPI Math Kernel Library. Intel Corporation. 2020. Available online: https://software.intel.com/content/www/us/en/develop/tools/oneapi/components/onemkl.html (accessed on 14 September 2022).
- NVIDIA Corporation. CUDA: Compute Unified Device Architecture. NVIDIA Corporation. 2022. Available online: http://developer.nvidia.com/object/cuda.html (accessed on 15 September 2022).
- Iles, G.; Jones, J.; Rose, A. Experience powering Xilinx Virtex-7 FPGAs. J. Instrum. 2013, 8, 12037. [Google Scholar] [CrossRef]
- Wang, K.; Huo, S.; Liu, B.; Wang, Z.; Ren, T. An Adaptive Low Computational Cost Alternating Direction Method of Multiplier for RELM Large-Scale Distributed Optimization. Mathematics 2024, 12, 43. [Google Scholar] [CrossRef]
- Jagadeesan, J.; Subashree, D.; Kirupanithi, D.N. An Optimized Ensemble Support Vector Machine-Based Extreme Learning Model for Real-Time Big Data Analytics and Disaster Prediction. Cogn. Comput. 2023, 15, 2152–2174. [Google Scholar] [CrossRef]
- Wang, Z.; Huo, S.; Xiong, X.; Wang, K.; Liu, B. A Maximally Split and Adaptive Relaxed Alternating Direction Method of Multipliers for Regularized Extreme Learning Machines. Mathematics 2023, 11, 3198. [Google Scholar] [CrossRef]
- Wang, G.; Soo, Z.S.D. BE-ELM: Biological ensemble Extreme Learning Machine without the need of explicit aggregation. Expert Syst. Appl. 2023, 230, 120677. [Google Scholar] [CrossRef]
- Zhang, Y.; Dai, Y.; Wu, Q. A novel regularization paradigm for the extreme learning machine. Neural Process. Lett. 2023, 55, 7009–7033. [Google Scholar] [CrossRef]
- Gelvez-Almeida, E.; Barrientos, R.J.; Vilches-Ponce, K.; Mora, M. Parallel training of a set of online sequential extreme learning machines. In Proceedings of the 2022 41st International Conference of the Chilean Computer Science Society (SCCC), Santiago, Chile, 21–25 November 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–4. [Google Scholar] [CrossRef]
- Gelvez-Almeida, E.; Barrientos, R.J.; Vilches-Ponce, K.; Mora, M. Parallel model of online sequential extreme learning machines for classification problems with large-scale databases. In Proceedings of the XI Jornadas de Cloud Computing, Big Data & Emerging Topics, Universidad de la Plata, La Plata, Argentina, 27–29 June 2023. [Google Scholar]
- Chidambaram, S.; Gowthul Alam, M. An Integration of Archerfish Hunter Spotted Hyena Optimization and Improved ELM Classifier for Multicollinear Big Data Classification Tasks. Neural Process. Lett. 2022, 54, 2049–2077. [Google Scholar] [CrossRef]
- Hira, S.; Bai, A. A Novel MapReduced Based Parallel Feature Selection and Extreme Learning for Micro Array Cancer Data Classification. Wirel. Pers. Commun. 2022, 123, 1483–1505. [Google Scholar] [CrossRef]
- Rajpal, S.; Agarwal, M.; Rajpal, A.; Lakhyani, N.; Saggar, A.; Kumar, N. COV-ELM classifier: An Extreme Learning Machine based identification of COVID-19 using Chest X-Ray Images. Intell. Decis. Technol. 2022, 16, 193–203. [Google Scholar] [CrossRef]
- Zha, L.; Ma, K.; Li, G.; Fang, Q.; Hu, X. A robust double-parallel extreme learning machine based on an improved M-estimation algorithm. Adv. Eng. Inform. 2022, 52, 101606. [Google Scholar] [CrossRef]
- Vidhya, M.; Aji, S. Parallelized extreme learning machine for online data classification. Appl. Intell. 2022, 52. [Google Scholar] [CrossRef]
- Rath, S.; Tripathy, A.; Swagatika, S. Application of ELM-mapreduce technique in stock market forecasting. In Intelligent and Cloud Computing; Mishra, D., Buyya, R., Mohapatra, P., Patnaik, S., Eds.; Springer: Singapore, 2021; Volume 2, pp. 469–476. [Google Scholar] [CrossRef]
- Ji, H.; Wu, G.; Wang, G. Accelerating ELM training over data streams. Int. J. Mach. Learn. Cybern. 2021, 12, 87–102. [Google Scholar] [CrossRef]
- Luo, F.; Liu, G.; Guo, W.; Chen, G.; Xiong, N. ML-KELM: A Kernel Extreme Learning Machine Scheme for Multi-Label Classification of Real Time Data Stream in SIoT. IEEE Trans. Netw. Sci. Eng. 2021, 9, 1–12. [Google Scholar] [CrossRef]
- Tahir, G.A.; Loo, C.K. Progressive kernel extreme learning machine for food image analysis via optimal features from quality resilient CNN. Appl. Sci. 2021, 11, 9562. [Google Scholar] [CrossRef]
- Dong, Z.; Lai, C.S.; Zhang, Z.; Qi, D.; Gao, M.; Duan, S. Neuromorphic extreme learning machines with bimodal memristive synapses. Neurocomputing 2021, 453, 38–49. [Google Scholar] [CrossRef]
- Ezemobi, E.; Tonoli, A.; Silvagni, M. Battery State of Health Estimation with Improved Generalization Using Parallel Layer Extreme Learning Machine. Energies 2021, 14, 2243. [Google Scholar] [CrossRef]
- Xu, Y.; Liu, H.; Long, Z. A distributed computing framework for wind speed big data forecasting on Apache Spark. Sustain. Energy Technol. Assess. 2020, 37, 100582. [Google Scholar] [CrossRef]
- Li, X.; Liu, J.; Niu, P. Least Square Parallel Extreme Learning Machine for Modeling NOx Emission of a 300MW Circulating Fluidized Bed Boiler. IEEE Access 2020, 8, 79619–79636. [Google Scholar] [CrossRef]
- Liang, Q.; Long, J.; Coppola, G.; Zhang, D.; Sun, W. Novel decoupling algorithm based on parallel voltage extreme learning machine (PV-ELM) for six-axis F/M sensors. Robot.-Comput.-Integr. Manuf. 2019, 57, 303–314. [Google Scholar] [CrossRef]
- Dokeroglu, T.; Sevinc, E. Evolutionary parallel extreme learning machines for the data classification problem. Comput. Ind. Eng. 2019, 130, 237–249. [Google Scholar] [CrossRef]
- Dean, J.; Ghemawat, S. MapReduce: Simplified data processing on large clusters. In Proceedings of the 6th Symposium on Operating Systems Design and Implementation. USENIX Association, San Francisco, CA, USA, 6–8 December 2004; Volume 6, pp. 137–149. [Google Scholar]
- Dean, J.; Ghemawat, S. MapReduce: Simplified data processing on large clusters. Commun. ACM 2008, 51, 107–113. [Google Scholar] [CrossRef]
- Gayathri, T.; Bhaskari, D.L. Oppositional Cuckoo Search Optimization based Clustering with Classification Model for Big Data Analytics in Healthcare Environment. J. Appl. Sci. Eng. 2021, 25, 743–751. [Google Scholar] [CrossRef]
- Yao, L.; Ge, Z. Distributed parallel deep learning of Hierarchical Extreme Learning Machine for multimode quality prediction with big process data. Eng. Appl. Artif. Intell. 2019, 81, 450–465. [Google Scholar] [CrossRef]
- Ku, J.; Zheng, B. Distributed extreme learning machine with kernels based on MapReduce for spectral-spatial classification of hyperspectral image. In Proceedings of the 2017 IEEE International Conference on Computational Science and Engineering (CSE) and IEEE International Conference on Embedded and Ubiquitous Computing (EUC), Guangzhou, China, 21–24 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 325–332. [Google Scholar] [CrossRef]
- Pang, J.; Gu, Y.; Xu, J.; Kong, X.; Yu, G. Parallel multi-graph classification using extreme learning machine and MapReduce. Neurocomputing 2017, 261, 171–183. [Google Scholar] [CrossRef]
- Inaba, F.K.; Salles, E.O.T.; Perron, S.; Caporossi, G. DGR-ELM–distributed generalized regularized ELM for classification. Neurocomputing 2018, 275, 1522–1530. [Google Scholar] [CrossRef]
- Huang, S.; Wang, B.; Qiu, J.; Yao, J.; Wang, G.; Yu, G. Parallel ensemble of online sequential extreme learning machine based on MapReduce. Neurocomputing 2016, 174, 352–367. [Google Scholar] [CrossRef]
- Wang, B.; Huang, S.; Qiu, J.; Liu, Y.; Wang, G. Parallel online sequential extreme learning machine based on MapReduce. Neurocomputing 2015, 149, 224–232. [Google Scholar] [CrossRef]
- Bi, X.; Zhao, X.; Wang, G.; Zhang, P.; Wang, C. Distributed Extreme Learning Machine with kernels based on MapReduce. Neurocomputing 2015, 149, 456–463. [Google Scholar] [CrossRef]
- Han, D.H.; Zhang, X.; Wang, G.R. Classifying Uncertain and Evolving Data Streams with Distributed Extreme Learning Machine. J. Comput. Sci. Technol. 2015, 30, 874–887. [Google Scholar] [CrossRef]
- Xiang, J.; Westerlund, M.; Sovilj, D.; Pulkkis, G. Using extreme learning machine for intrusion detection in a big data environment. In Proceedings of the 2014 Workshop on Artificial Intelligent and Security Workshop, Association for Computing Machinery, Scottsdale, AZ, USA, 7 November 2014; pp. 73–82. [Google Scholar] [CrossRef]
- Xin, J.; Wang, Z.; Chen, C.; Ding, L.; Wang, G.; Zhao, Y. ELM*: Distributed extreme learning machine with MapReduce. World Wide Web 2014, 17, 1189–1204. [Google Scholar] [CrossRef]
- He, Q.; Shang, T.; Zhuang, F.; Shi, Z. Parallel extreme learning machine for regression based on MapReduce. Neurocomputing 2013, 102, 52–58. [Google Scholar] [CrossRef]
- Zaharia, M.; Chowdhury, M.; Franklin, M.J.; Shenker, S.; Stoica, I. Spark: Cluster Computing with Working Sets. In Proceedings of the 2nd USENIX Conference on Hot Topics in Cloud Computing, USENIX Association, Boston, MA, USA, 22–25 June 2010; pp. 1–10. [Google Scholar]
- Jaya Lakshmi, A.; Venkatramaphanikumar, S.; Venkata, K.K.K. Prediction of Cardiovascular Risk Using Extreme Learning Machine-Tree Classifier on Apache Spark Cluster. Recent Adv. Comput. Sci. Commun. 2022, 15, 443–455. [Google Scholar] [CrossRef]
- Kozik, R.; Choraś, M.; Ficco, M.; Palmieri, F. A scalable distributed machine learning approach for attack detection in edge computing environments. J. Parallel Distrib. Comput. 2018, 119, 18–26. [Google Scholar] [CrossRef]
- Kozik, R. Distributing extreme learning machines with Apache Spark for NetFlow-based malware activity detection. Pattern Recognit. Lett. 2018, 101, 14–20. [Google Scholar] [CrossRef]
- Oneto, L.; Fumeo, E.; Clerico, G.; Canepa, R.; Papa, F.; Dambra, C.; Mazzino, N.; Anguita, D. Dynamic Delay Predictions for Large-Scale Railway Networks: Deep and Shallow Extreme Learning Machines Tuned via Thresholdout. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 2754–2767. [Google Scholar] [CrossRef]
- Oneto, L.; Fumeo, E.; Clerico, G.; Canepa, R.; Papa, F.; Dambra, C.; Mazzino, N.; Anguita, D. Train Delay Prediction Systems: A Big Data Analytics Perspective. Big Data Res. 2018, 11, 54–64. [Google Scholar] [CrossRef]
- Duan, M.; Li, K.; Liao, X.; Li, K. A Parallel Multiclassification Algorithm for Big Data Using an Extreme Learning Machine. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 2337–2351. [Google Scholar] [CrossRef]
- Liu, T.; Fang, Z.; Zhao, C.; Zhou, Y. Parallelization of a series of extreme learning machine algorithms based on Spark. In Proceedings of the 2016 IEEE/ACIS 15th International Conference on Computer and Information Science (ICIS), IEEE, Okayama, Japan, 26–29 June 2016; pp. 1–5. [Google Scholar] [CrossRef]
- Navarro, C.A.; Carrasco, R.; Barrientos, R.J.; Riquelme, J.A.; Vega, R. GPU Tensor Cores for Fast Arithmetic Reductions. IEEE Trans. Parallel Distrib. Syst. 2021, 32, 72–84. [Google Scholar] [CrossRef]
- Hou, X.C.; Lai, X.P.; Cao, J.W. A Maximally Split Generalized ADMM for Regularized Extreme Learning Machines. Tien Tzu Hsueh Pao/Acta Electron. Sin. 2021, 49, 625–630. [Google Scholar] [CrossRef]
- El Zini, J.; Rizk, Y.; Awad, M. An optimized parallel implementation of non-iteratively trained recurrent neural networks. J. Artif. Intell. Soft Comput. Res. 2021, 11, 33–50. [Google Scholar] [CrossRef]
- Li, S.; Niu, X.; Dou, Y.; Lv, Q.; Wang, Y. Heterogeneous blocked CPU-GPU accelerate scheme for large scale extreme learning machine. Neurocomputing 2017, 261, 153–163. [Google Scholar] [CrossRef]
- Chen, C.; Li, K.; Ouyang, A.; Tang, Z.; Li, K. GPU-Accelerated Parallel Hierarchical Extreme Learning Machine on Flink for Big Data. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 2740–2753. [Google Scholar] [CrossRef]
- Lam, D.; Wunsch, D. Unsupervised Feature Learning Classification With Radial Basis Function Extreme Learning Machine Using Graphic Processors. IEEE Trans. Cybern. 2016, 47, 224–231. [Google Scholar] [CrossRef]
- Van Heeswijk, M.; Miche, Y.; Oja, E.; Lendasse, A. GPU-accelerated and parallelized ELM ensembles for large-scale regression. Neurocomputing 2011, 74, 2430–2437. [Google Scholar] [CrossRef]
- Jezowicz, T.; Gajdo, P.; Uher, V.; Snáel, V. Classification with extreme learning machine on GPU. In Proceedings of the 2015 International Conference on Intelligent Networking and Collaborative Systems, Taipei, Taiwan, 2–4 September 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 116–122. [Google Scholar] [CrossRef]
- Li, J.; Guo, B.; Shen, Y.; Li, D.; Wang, J.; Huang, Y.; Li, Q. GPU-memory coordinated energy saving approach based on extreme learning machine. In Proceedings of the 2015 IEEE 17th International Conference on High Performance Computing and Communications, 2015 IEEE 7th International Symposium on Cyberspace Safety and Security, and 2015 IEEE 12th International Conference on Embedded Software and Systems, New York, NY, USA, 24–26 August 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 827–830. [Google Scholar] [CrossRef]
- Krawczyk, B. GPU-Accelerated Extreme Learning Machines for Imbalanced Data Streams with Concept Drift. Procedia Comput. Sci. 2016, 80, 1692–1701. [Google Scholar] [CrossRef]
- Dwivedi, S.; Vardhan, M.; Tripathi, S. Multi-Parallel Adaptive Grasshopper Optimization Technique for Detecting Anonymous Attacks in Wireless Networks. Wirel. Pers. Commun. 2021, 119, 2787–2816. [Google Scholar] [CrossRef]
- Li, Y.; Zhang, S.; Yin, Y.; Xiao, W.; Zhang, J. Parallel one-class extreme learning machine for imbalance learning based on Bayesian approach. J. Ambient. Intell. Humaniz. Comput. 2024, 15, 1745–1762. [Google Scholar] [CrossRef]
- Ming, Y.; Zhu, E.; Wang, M.; Ye, Y.; Liu, X.; Yin, J. DMP-ELMs: Data and model parallel extreme learning machines for large-scale learning tasks. Neurocomputing 2018, 320, 85–97. [Google Scholar] [CrossRef]
- Henríquez, P.A.; Ruz, G.A. Extreme learning machine with a deterministic assignment of hidden weights in two parallel layers. Neurocomputing 2017, 226, 109–116. [Google Scholar] [CrossRef]
- Luo, M.; Zhang, L.; Liu, J.; Guo, J.; Zheng, Q. Distributed extreme learning machine with alternating direction method of multiplier. Neurocomputing 2017, 261, 164–170. [Google Scholar] [CrossRef]
- Wang, Y.; Dou, Y.; Liu, X.; Lei, Y. PR-ELM: Parallel regularized extreme learning machine based on cluster. Neurocomputing 2016, 173, 1073–1081. [Google Scholar] [CrossRef]
Reference | Type | Focus | Advantages/Disadvantages | No. Citations * | ||
---|---|---|---|---|---|---|
WoS | Scopus | Scholar | ||||
Huérfano-Maldonado et al. (2023) [11] | Journal: Neurocomputing | Medical image processing using ELM | ELM offers advantages such as fewer training parameters, fast learning speed, and high generalization ability. However, most publications focus on supervised learning, with fewer on unsupervised learning. | 3 | 4 | 5 |
Patil and Sharma (2023) [12] | Conf: Int. Conf. on Computational Intelligence and Sustainable Engineering Solutions | Theories, algorithms, and applications of ELM | ELM exhibits fast training and efficient processing of large data sets; however, it is constrained by its limited capacity to learn complex patterns. | - | 0 | 0 |
Kaur et al. (2023) [13] | Journal: Multimedia Tools and Applications | Multilayer ELM | The applications of ML-ELM in parallel and distributed computing are open. Also, its effectiveness for big data applications can be investigated further. | 0 | 3 | 3 |
Vásquez-Coronel et al. (2023) [14] | Journal: Artificial Intelligence Review | Multilayer ELM | With the advancement of technology and the Internet, the amount of data continues to increase exponentially, giving rise to a current open problem in machine learning. | 4 | 4 | 5 |
Wang et al. (2022) [15] | Journal: Multimed. Tools Appl. | ELM neural network | ELM has the potential of playing a more important role in big data. | 120 | 138 | 241 |
Zheng et al. (2022) [16] | Journal: Big Data Research | Data-stream classification based on ELM | ELM-based algorithms have better generalization ability and less computation time. However, these algorithms are little focused on multi-label data-stream classification. | 3 | 6 | 7 |
Martínez et al. (2022) [17] | Conf: Colombian Conf. Communications and Computing | Identification and classification of fingerprint databases | ELM-based algorithms are an economical and accessible alternative to reduce the penetration rate in fingerprint databases compared to traditional algorithms. | - | 0 | 1 |
Kaur et al. (2022) [18] | Conf: Int. Conf. Machine Learning, Computer Systems, and Security | ELM on TF-IDF features for sentiment analysis | The need for a high number of hidden nodes in ELM is one of the drawbacks which needs to be addressed. | - | 0 | 0 |
Nilesh and Sunil (2021) [19] | Journal: EAI Endorsed Transactions on Industrial Networks and Intelligent Systems | Optimization algorithms for ELM | The efficiency, accuracy, and easy implementation in various fields are advantages. However, ELM cannot handle massive high-dimensional information. It needs additional hidden nodes. | - | 2 | 5 |
Mujal et al. (2021) [20] | Journal: Advanced Quantum Technologies | Quantum reservoir computing (QRC) and quantum ELM (QELM) | QRC and QELM are still taking their first steps, so it is premature to make quantitative comparisons with their much more advanced classical counterparts. | 33 | 41 | 79 |
Nilesh and Sunil (2021) [21] | Conf.: Int. Conf. Advanced Computing and Communication Systems | Optimization algorithm for ELM | Parallel and transmitted processing of ELM will become the following focal point. Identifying the appropriate number of hidden layer neurons is a disadvantage. | - | 7 | 9 |
Rodrigues et al. (2021) [22] | Journal: Informatics | Convolutional ELM (CELM) | The training time, test time, and accuracy are some advantages of CELM. However, when the number of layers increases, problems with increasing training time and a loss of generalization capacity emerge. | 8 | 15 | 19 |
Saldaña et al. (2021) [23] | Conf.: Brazilian Technology Symposium | ELM for business sales forecasts | There exists little information on the subject because it is a new topic. | - | 3 | 5 |
Wang et al. (2020) [24] | Journal: IEEE Access | Application of ELM in computer-aided diagnosis (CAD) | The research in this field has important medical and social value. ELM has a short processing time and also has good generalization performance. | 9 | 11 | 12 |
Wang et al. (2020) [25] | Journal: IEEE Access | Distributed and parallel ELM for big data | The limitation of hardware is a disadvantage of distributed ELM. Distributed ELM does not apply well to specific problems. | 4 | 7 | 9 |
Alaba et al. (2019) [26] | Journal: Neurocomputing | Advances and drawbacks of ELM | The adoption of parallel computing based on MapReduce and GPU acceleration has demonstrated better efficiency. | 38 | 44 | 51 |
Yibo et al. (2019) [27] | Journal: J. Phys.: Conf. Ser. | Prediction model of ELM | A combination forecasting method is better than a single forecasting method in forecasting accuracy. | - | 8 | 10 |
Li et al. (2019) [28] | Journal: Multimed. Tools Appl. | Improved ELM algorithms used for data stream | There is no ideal solution for determining the optimal number of hidden layer nodes. The application of improved ELM for data-stream classification is limited. | 16 | 22 | 25 |
Eshtay et al. (2019) [29] | Journal: Int. J. Machi. Learn. and Cyber. | ELM based on metaheuristics | In real life specific problems are still at an early stage. There is still no work that has researched the performance of this type of model using large-scale data sets. | 33 | 38 | 47 |
Ghosh et al. (2018) [30] | Conf.: Int. Conf. Recent Trends in Image Processing and Pattern Recognition | ELM and the evolution of its variants | ELM is well-defined in the field of pattern recognition, medical diagnosis, and forecasting areas. The applications of ELM are still open for parallel and distributed computing. | - | 5 | 7 |
Zhang et al. (2018) [31] | Conf.: Int. Conf. Control, Decision and Information Technologies (CoDIT) | Online sequential extreme learning machine (OS-ELM) | OS-ELM is a faster and more accurate algorithm as compared to other online learning algorithms. The improved OS-ELM algorithms need to be network structure adjusted to improve learning prominence. | - | 7 | 10 |
Alade et al. (2018) [32] | Conf.: Int. Conf. Reliable Information and Communication Technology | Advances in ELM techniques and their applications | The relevance of ELM in artificial intelligence has brought about great interest. The handling of high-dimensional data is a disadvantage. | - | 19 | 26 |
Salaken et al. (2017) [33] | Journal: Neurocomputing | Transfer learning (TL) using ELM | The papers published to validate the effectiveness of ELM to solve transfer learning problems. However, these articles failed to compare their performance against existing algorithms. | 65 | 78 | 98 |
Albadra and Tiun (2017) [34] | Journal: Int. J. Appl. Engineering Research | Advances in ELM techniques and their applications | ELM’s major strength is that the learning parameters do not have to be iterative. However, the classification boundary of the parameters may not be optimal because they remain the same during training. | - | 86 | 111 |
Ali and Zolkipli (2016) [35] | Journal: ARPN Journal of Engineering and Applied Sciences | Integration of genetic algorithms (GAs) and ELM to function as an intrusion detection system (IDS) | Advantages of ELMs include ease of implementation and the ability to perform multi-class classification directly without using binary classification techniques in succession. | - | 9 | 12 |
Huanh et al. (2015) [36] | Journal: Neural Networks | Trends in ELM | ELM and its variants are efficient, accurate, and easy to implement. High-dimensional data analysis is a challenging problem for ELM and its variants. | 1240 | 1461 | 1839 |
Cao and Lin (2015) [37] | Journal: Mathematical Problems in Engineering | ELM on high-dimensional and large data applications | ELM has a fast data learning speed and easy implementation. Designing real-time processing systems and devices for applications is highly desired. | 33 | 66 | 116 |
Ding et al. (2015) [38] | Journal: Artif. Intell. Rev. | ELM algorithms, theory, and applications | ELM generates a unique optimal solution with the advantages of fast learning speed and generalization performance. However, setting the number of nodes in the hidden layer is a disadvantage. | 344 | 440 | 574 |
Deng et al. (2015) [39] | Journal: Sci. China Inf. Sci. | New trends and applications of ELM | Fast learning speed, ease of implementation, and minimal human intervention are significant advantages of ELM compared with traditional NNs and SVM. | 103 | 123 | 154 |
Ding et al. (2014) [40] | Journal: Neural Computing and Applications | ELM and its applications | The generalization performance of ELM turns out to be stable. Parallel and distributed computing of ELM are some open problems. | 163 | 213 | 283 |
Reference | Application | Architecture | Parallel Tool | Data | Results | Limitations |
---|---|---|---|---|---|---|
Li (2024) [62] | Sparse Triangular Solver (SpTRSV) | Sunway many-core processors. | swSparse library | 949 real square matrices and 32 complex square matrices from the SuiteSparse matrix data set. | Outperforms cuSparse on NVIDIA V100 GPUs and Intel MKL library, with better speedup for larger matrices (size > 10,000). | Further comparison is needed to determine how closely the observed performance matches the hardware peak. |
Gelvez-Almeida et al. (2023) [63] | Strassen algorithm | Server, 2 × Intel Xeon Gold 6238R (56 cores), 128 GB RAM. | OpenMP | Full-rank matrices with , and to . | Superior computation time for the Moore–Penrose generalized inverse compared to previously reported algorithms. | High memory consumption. |
Lukyanenko (2023) [64] | Conjugate gradient method | Supercomputer, Intel Xeon E5-2697 (14 cores) each node, 64 GB RAM, Tesla K40s GPU 11.56 GB. | MPI | Matrices of size with and . | The implementation provides a suitable approximate solution without increasing computational complexity. | The matrix dimensions are limited by the available memory. |
Suzuki et al. (2023) [65] | Blocks into ILU preconditioning (ILUB) | Server, 2 × Intel Xeon Gold 6148 (20 cores/CPU), 384 GB RAM; Intel Xeon Phi 7250 (68 cores), 96 GB RAM. | OpenMP | Several data sets of dimensions to . | Outperformed conventional ILU(0) preconditioning. | Future research should explore how ILUB performs when combined with others reordering techniques. |
Sabelfeld et al. (2023) [66] | Random vector estimator | Cluster, 2 × Intel Xeon E5-2697A v4 (32 cores, 64 threads per node), 128 GB RAM. | MPI and OpenMP | Dense matrices of size and . | The "RAM" implementation is faster and preferable for large linear systems if the matrix fits in the node’s RAM. | The implementations developed are restricted in problem size by the node’s limited memory. |
Catalán et al. (2022) [67] | QR and SVD factorization | Servers, Intel Xeon Gold 6138 (20 cores), 96 GB RAM; AMD EPYC 7742 (64 cores), 512 GB RAM. | OpenMP | Square matrices of dimensions 1000 to 16,000. | Outperforms Intel’s MKL and the AMD AOCL routine. It is highly competitive with PLASMA and outperforms its counterparts. | Manufacturers’ effort in tuning the performance of linear algebra libraries is rarely shared with the scientific community. |
Rivera et al. (2022) [68] | Barrett reduction for Wiedemann | Multicore CPU servers, Supercomputer ABACUS, NVIDIA TITAN. | OpenMP and CUDA | Quadratic, cubic, and quartic matrices of dimensions 55,000 to 266,000. | GPU is faster than the CPU for the cubic and quartic families. CPU outperforms GPU for the quadratic matrix. | The large size of this matrix causes a costly divergence between the thread computations. |
Li and Han (2022) [69] | Gauss–Newton method | PC, Intel i5-6400 2.7 GHz Quad Core processor, 16 GB RAM. | MATLAB | Unbalanced IEEE 33 and 123-bus systems. | The algorithm is computationally efficient and requires fewer iterations. | Each iteration requires solving a large linear system. |
Hwang et al. (2022) [70] | Gauss–Seidel method | Multiple input multiple output (MIMO) systems. | Rayleigh flat fading channel | Matrices with independent Gaussian random variables. | Up to a certain point, the performance of this model is better than the conventional Gauss–Seidel. | Performance with too much parallel computation is poorer than the Gauss–Seidel. |
Catalán et al. (2021) [71] | Gauss–Jordan elimination | Server, Intel Xeon Gold 6138 (20 cores), 96 GB RAM. | OpenMP | Square matrices of dimension n up to 30,000. | Yields a high core occupation, and its parallel performance exceeds that of IntelMKL, BLAS and PLASMA. | Non-overlapping execution of the cores is likely to result in poor performance. |
Marrakchi and Jemni (2021) [72] | Gaussian elimination | Cluster Econome, Intel Xeon E5-2660 (8 cores). | OpenMP | Square matrices of dimensions 1000 to 3500. | A higher degree of parallelism is achieved, and the approach is adjusted to theoretical values. | Efficiency decreases with each additional core. |
Lu et al. (2021) [73] | LU and Cholesky factorizations | MLU270-S4 AI card (GDRAM and NRAM). | BANG C language | Square matrices of dimensions 128 to 8192. | A variety of optimizations demonstrated their effectiveness. | When the size of the matrix is larger, data transmission congestion occurs. |
Lee and Achar (2021) [74] | LU factorization | 8 × Intel Xeon i7-9700F, 16 GB RAM, NVIDIA Turing RTX2060 (1920 cores), 6 GB DRAM. | CUDA | Several matrices of dimensions 1879 to 1,585,478. | The proposed advancements provide superior performance compared to GLU 3.0 and KLU. | Because the GPU has limited global memory, it is necessary to limit the number of columns used at one time for parallel processing. |
Zhang et al. (2021) [75] | LDL factorization | Xilinx Virtex-7 XC7VX690T FPGA. | Xilinx Vivado toolset | Square and triangular matrices of dimensions 32 to 128. | Square matrix inversion was achieved with different sizes in the FPGA chip. | Matrix inversion in the different FPGAs is not discussed. |
Rubensson et al. (2021) [76] | Localized inverse factorization | Rackham cluster, Intel Xeon E5-2630 v4 (10 cores), 128 GB RAM. | MPI and OpenBLAS | Several matrices of dimensions up to 5,373,954. | Represents a dramatic improvement over the regular recursive inverse factorization. | Localized inverse factorization is completely dominated by the solutions to the subproblems. |
Rodriguez et al. (2021) [77] | QR factorization | Xilinx VU7P FPGA and Xilinx VU9P FPGA. | LUT RAM | Square matrices of dimensions 4000 to 131,000. | This design outperforms highly optimized QR solvers running on CPUs and GPUs. | Overall performance is limited when the input matrix has few columns. |
Duan and Dinavahi (2021) [78] | Linking-domain extraction (LDE)-based decomposition method | Xilinx VCU-118 board with the XCVU9P FPGA at 100 MHz frequency and NVIDIA Tesla V100 GPU with 5012 cores. | Not reported | Square matrices of dimensions 30 to 301. | LDE method can compute the matrix inversion directly and can also run faster compared to the Schur complement. | The connections between interface nodes are not dense, and there is no trans-conductance between the interface nodes and the other nodes. |
Shäfer et al. (2021) [79] | Cholesky factorization | Server, Intel Skylake 2.10 GHz (32 threads), 192 GB RAM; PC, Intel Core i7-6400 4.00 GHz, 64 GB RAM. | IntelMKL | Square matrices of dimensions 10,000 to 1,000,000. | The optimal inverse Cholesky factor of a positive-definite matrix, subject to a sparsity pattern, can be computed in closed form. | More efficient squaring rules need to be implemented to compete with the state of the art in terms of wall clock times. |
Boffi et al. (2021) [80] | Iterative incomplete LU (ILU) method | Laptop, Intel i7-6700HQ, 16 GB RAM. | MATLAB | Square matrices of dimensions 10,000 to 1,000,000. | The new preconditions have been successfully applied to linear systems and eigenvalue problems. | The proposed algorithms may become unstable and fail for some matrices. |
Ahmadi et al. (2021) [81] | Jacobi-embedded Gauss–Seidel method | Server, 2 × Intel Xeon Gold 6148 (20 cores/CPU), 4 × NVIDIA Tesla V100-SXM2-16 GB. | MATLAB | Square matrices of dimensions 500 to 25,000. | Performance up to 7× faster on multicore CPUs and 87× on many-core GPUs. | Limitations with MATLAB linear algebra libraries for GPU implementation. |
Liu et al. (2020) [82] | Blocked adaptive cross approximation (BACA) and SVD | Cori Haswell at NERSC (2388 dual-socket nodes), Intel Xeon E5-2698v3 (16 cores/each), 128 GB DDR4. | PBLAS and ScaLAPACK | Several matrices of dimensions 1000 to 21,788. | Robustness and favorable parallel performance compared to the baseline ACA algorithm. | The effects of block size variation merit further analysis. |
Davis et al. (2020) [83] | LU factorization | Supercomputer (HPC2N), Intel Xeon E5-2690v4 (2×14 cores), 128 GB RAM. | OpenMP | Square matrices of dimensions 53,000 to 659,000. | Excellent performance for highly unsymmetrical matrices. | Sometimes produces factorizations with more fill-in. |
Yang et al. (2020) [84] | Gauss–Seidel method | PC, Intel Core i5. | MATLAB (R2016a) | Matrices of dimensions 20 to 1000. | Numerical results show that the method is valid. | The dimensions of the matrices are not large enough. |
Li and Zhang (2020) [85] | Gauss–Seidel method | Cluster, CPU IBM POWER, NVIDIA P100 and V100. | CUDA | Laplacian and general matrices with different dimensions. | Efficient algorithm compared to state-of-the-art software packages. | Performance deteriorates with the amount of fill-in in the ILU factorizations. |
Singh et al. (2020) [86] | Gauss–Newton method | Blue Waters and Stampede2 Supercomputers of Texas Advanced Computing Center. | Intel compilers, MKL library for BLAS, and Cyclops | Square matrices with different dimensions. | Good scalability for the implementation of the Gauss–Newton method. | The method does not apply to triangular matrices. |
Alyahya et al. (2020) [87] | Jacobi method | Aziz Supercomputer of King Abdulaziz University, Jeddah. | Intel MIC and OpenMP | Sparse matrices with 28M rows and 640M non-zero elements. | Speedup to 27.75× compared to the sequential method. | A method is needed for larger sparse SLEs of various application domains. |
Huang et al. (2020) [88] | Jacobi, SOR, and other iterative methods | Server, 2 × Intel Xeon E5-2640 v3 2.60 GHz (16 cores). | OpenMP | Three-dimensional sphere DDA (SDDA). 10,000 spheres and 200,000 calculation steps. | About 6× faster than serial computing. | Other approaches can be used to improve the efficiency of solving the equations in the DDA. |
Reference | Network Type | MPGI | Orthogonal Projection | SVD | Cholesky | Iterative Methods | Not Reported |
---|---|---|---|---|---|---|---|
Wang et al. (2024) [97] | Regularized ELM | ✓ | |||||
Jagadeesa et al. (2023) [98] | SVM-Based ELM | ✓ | |||||
Wang et al. (2023) [99] | Regularized ELM | ✓ | |||||
Wang and Soo (2023) [100] | Ensemble ELM | ✓ | ✓ | ||||
Zhang et al. (2023) [101] | Regularized ELM | ✓ | |||||
Gelvez-Almeida et al. (2022–2023) [102,103] | Ensemble Online ELM | ✓ | ✓ | ||||
Polat and Kayhan (2022) [51] | Online ELM | ✓ | |||||
Chidambaram and Gowthul (2022) [104] | Improved ELM | ✓ | |||||
Hira and Bai (2022) [105] | Regularized ELM | ✓ | |||||
Rajpal et al. (2022) [106] | Standard ELM | ✓ | |||||
Zha et al. (2022) [107] | Robust ELM | ✓ | ✓ | ||||
Vidhya and Aji (2022) [108] | Online ELM | ✓ | ✓ | ||||
Zehai et al. (2021) [50] | Ensemble Online ELM | ✓ | ✓ | ||||
Wu et al. (2021) [55] | Kernel ELM | ✓ | |||||
Rath et al. (2021) [109] | Hierarchical ELM | ✓ | |||||
Ji et al. (2021) [110] | Online ELM | ✓ | |||||
Luo et al. (2021) [111] | Kernel ELM | ✓ | |||||
Tahir and Loo (2021) [112] | Kernel ELM | ✓ | |||||
Dong et al. (2021) [113] | Standard ELM | ✓ | ✓ | ||||
Ezemobi et al. (2021) [114] | Deterministic ELM | ✓ | |||||
Xu et al. (2020) [115] | Distributed | ✓ | |||||
Li et al. (2020) [116] | Online ELM | ✓ | |||||
Safaei et al. (2019) [48] | Online ELM | ✓ | |||||
Grim et al. (2019) [49] | Online ELM | ✓ | |||||
Liang et al. (2019) [117] | Voltage ELM | ✓ | |||||
Dokeroglu and Sevinc (2019) [118] | Evolutionary ELM | ✓ |
Reference | Application | Architecture | Data | Size (GB) * | Limitations |
---|---|---|---|---|---|
Chidambaram and Gowthul (2022) [104] | Classification | PC, Intel Core i5 9th Gen, 8 GB RAM. | Rotten Tomatoes movie and critic review data set (DS-I) and dermatology data set (DS-II). | 0.004 | A basic architecture was used during the experiments. |
Hira and Bai (2022) [105] | Classification | Different cluster managers. | GSE13159, SRBCT, acute leukaemia, leukaemia 2, and colon cancer. | Information not provided | MapReduce is not suitable for applications that share data in more than one step. |
Gayathri et al. (2021) [121] | Classification | Hadoop clusters. | PIMA Indians diabetes and activity recognition data set. | 0.00005; 0.005 | Feature selection and dimensionality reduction techniques have not been considered. |
Rath et al. (2021) [109] | Regression | Not reported. | NASDAQ (ten years of historical data). | Information not provided | It must wait for all the parallel work tasks to be completed before moving on to the next step. |
Yao and Ge (2019) [122] | Regression | Cluster (4 PCs), Intel Core i5-4590 3.30 GHz, 4 GB RAM. | DCS (110,000 samples × 20 processes). | Information not provided. | The files must have the same size. |
Ku and Zheng (2017) [123] | Classification | Not reported. | AVIRIS Indian Pines (10,366 samples), ROSIS Pavia University (46,697 samples). | Information not provided | A second computer is needed when training data overcome some limitations when calculating the matrix and pseudo-inverse. |
Pang et al. (2017) [124] | Classification | Cluster (31 PCs), 2 × 3.1 GHz CPUs, 8 GB RAM, 500 GB hard disk. | DBLP (1,817 AI y 1,817 CV); Synthetic (500,000 pos. and 500,000 neg.; 4,997,537 total). | Information not provided | The proposed algorithm is time-consuming because the algorithm consists of several MapReduce jobs. |
Inaba et al. (2016) [125] | Classification | PC, Intel Core i7 3.6 GHz, 8 GB RAM. | Various. DNA (3186 × 180); Satimage (6435 × 36). | 0.004; 0.002 | Training time for DGR-ELM can be improved by using GPU-based linear algebra packages. |
Huang et al. (2016) [126] | Classification | Cluster (5 PCs), 2 × Intel Xeon E5-2620 (6 cores), 32 GB RAM. | Various. KDDcup99 (5,190,731 × 41); Synthetic (5,120,000 × 512). | 0.67; 19.53 | More scheduling and ensemble methods need to be integrated to make the algorithm more suitable for heterogeneous environments. |
Wang et al. (2015) [127] | Classification | Cluster (9 PCs), Intel Quad core Q8400 2.66 GHz, 4 GB RAM. | Various. Covtype (580,000 × 54); Synthetic (1,280,000 × 128). | 0.23; 1.22 | The block size of the algorithm influences the learning performance. |
Bi et al. (2015) [128] | Classification | Cluster (9 PCs), Intel Quad Core 2.66 GHz, 4 GB RAM. | Various. Synthetic (25,000 × 5000). | 0.93 | Some matrix operations cannot be implemented on MapReduce directly. |
Han et al. (2015) [129] | Classification | Cluster (8 PCs), Intel core 2 Quad Q8400 2.66 GHz, 4 GB RAM. | Various. Iris (6,000,000 × 4); Spambase (18,400,000 × 57). | 0.18; 7.81 | When the block size is larger, the amount of computation increases, and the concurrency of the algorithm decreases. |
Xiang et al. (2014) [130] | Classification | PC, 2 × CPU 2.53 GHz (4 cores/CPU), 32 GB RAM. | KDDcup99 (1 to 3 million samples). | Information not provided | A distributed approach is needed to solve a linear system to enable the use of larger network architectures. |
Xin et al. (2014) [131] | Regression | Cluster (9 PCs), Intel Quad Core 2.66 GHz, 4 GB RAM. | Synthetic (3 to 7 million samples). | 1.4 to 3.27 | Training time increases with an increasing number of training lengths. |
He et al. (2013) [132] | Regression | Cluster (10 PCs), 4 × 2.8 GHz cores, 4 GB RAM. | Stock (950 × 12). | 0.000085 | It is necessary to make better use of IT resources. |
Reference | Application | Architecture | Data | Size (GB) * | Limitations |
---|---|---|---|---|---|
Jagadeesan et al. (2023) [98] | Classification | Cluster (3 nodes), Intel Core i5 3470 3.2 GHz (8 cores, 16 threads), 16 GB RAM. | Over 11,000 tweets relevant to disasters are included, along with tweets from news agencies and disaster relief organizations on Twitter. | Information not provided | Reliance on random operators for implementation, which can slightly affect output results when applied in different domains. |
Jaya et al. (2022) [134] | Classification | Cluster (8 nodes), 4 × 2.4 GHz processors, 200 GB hard disk, 16 GB RAM. | Various from UCI data repository. | Information not provided | Limited use of classifiers available in the distributed Spark environment. |
Ji et al. (2021) [110] | Classification | Cluster (5-node OMNISKY), 2 × Intel Xeon E5-2603V4 1.70 GHz, 2 × NVIDIA Quadro M4000, 16 TB hard disk, 64 GB RAM. | Synthetic data set. | 0.1 | The combination of distributed computing and GPU acceleration has not been used with a real data set. |
Luo et al. (2021) [111] | Classification | Single-server and multi-node in Spark cluster, 4-core 2 GHz CPU, 8 GB RAM. | Various from Mulan and extreme repository data sets. | Information not provided | The analysis of the relationship between the labels should be strengthened. |
Xu et al. (2020) [115] | Regression | Different nodes on the clusters of the Apache Spark platform. | Three real wind speed data sets and a group of analog wind speed big data (1000 to 600,000 samples). | Information not provided | The Spark platform needs time to start the framework and distribute tasks. |
Kozik et al. 2018 [135] | Classification | HPC cluster. | Real-world attacks data set. | Information not provided | The computation time increases as the number of training samples increases. |
Kozik (2018) [136] | Classification | Computing cluster. | CTU data set. | 5.2 to 60 | Other algorithms perform better in terms of accuracy and error rates. |
Oneto et al. (2017–2018) [137,138] | Regression | PC, 4 × Intel Xeon E5-4620 2.20 GHz, 128 GB RAM, 500 GB SSD disk; 4 × n1-standard-16 of the Google Compute Engine, 60 GB RAM, 16 cores, 500 GB SSD disk. | Real data of Rete Ferroviaria Italiana (RFI) and the Italian infrastructure manager (IM). | Information not provided | It does not take into account information available from external sources. |
Duan et al. (2017) [139] | Classification | Cluster (10–35 servers), 2.5 GHz Core, 16 GB RAM, 2 TB disk. | Various. Hypertension (38 M × 8); heart disease (10 M × 35). | 4.65; 5.08 | There is a high probability that one or more nodes will not be able to function. |
Liu et al. (2016) [140] | Classification | Local Cluster, VMare workstation 9.0.2 for Windows Ubuntu system. | Various. Forest types prediction (581,012 × 54). | 2.34 | As the number of hidden layer nodes increases, the learning and training time will gradually increase. |
Reference | Application | Architecture | Parallel Tool | Data | Size (GB) * | Limitations |
---|---|---|---|---|---|---|
Wang et al. (2023) [99] | Classification | PCs, Intel Core i7-10700 (8 cores) 2.9 GHz; Intel Core i7-4790 (4 cores) 3.60 GHz; Intel Core i7-8700 (6 cores) 3.2 GHz; NVIDIA GeForce GT 730 | MATLAB toolbox (gpuArray function). | Various. Gissete (7000 samples). | 0.26 | Requires a large amount of storage and computation. |
Polat and Kayhan (2022) [51] | Classification | PC, Intel Core i7-6700K, 32 GB RAM, NVIDIA GeForce GTX 1070, 1920 CUDA cores. | C++ and CUDA Toolkit 10.2. | Various. Landsat Satellite (6435 × 36); MNIST (70,000 × 784). | 0.002; 0.409 | Computationally, it is very time-consuming to apply cross-validation for each different hyper-parameter for online learning. |
Tahir and Loo (2021) [112] | Classification | Server, 16 GB GPU, 64 GB RAM. | Python Django framework. | Various. Food101 (101 images); VireoFood-172 (110,241 images). | 0.041; 7.22 | Various image quality distortions affect the robustness of the features. |
Hou et al. (2021) [142] | Classification | PC, Intel Core i7-8700K 3.7 GHz, NVIDIA GeForce RTX-2080Ti, 64 GB RAM. | MATLAB 2017. | Various. NORB (74,300 images); MNIST (70,000 × 784). | 3.34; 0.35 | Parameter selection and efficiency are not explored when memory is limited. |
El Zini et al. (2021) [143] | Regression | PCs, Intel core-i7 and Core-i5, NVIDIA Tesla K20m (2688 CUDA cores), NVIDIA Quadro K2000, 16 GB RAM. | CUDA. | Various. Electric Motor Temperature data set (998,000 instances). | Up to 0.12 | Portability and scalability of the proposed algorithm need to be further studied. |
Rajpal et al. (2021) [106] | Classification | NVIDIA Tesla K80 (Google Collaboratory). | Python. | Chest X-ray images (CXRs). | Information not provided | It is necessary to explore the segmentation of the pulmonary region. |
Grim et al. (2019) [49] | Regression | Virtual machine, 6 vCPUs Xeon E5-2690 2.60 GHz, 1 Tesla K80 823.5 MHz and 11.5 GB, 56 GB RAM. | OpenBLAS, IntelMKL and MAGMA. | Environmental data set with samples of particulate matter concentrations. | Information not provided | For a small set of samples, the number of function calls is higher and this time overhead becomes considerable. |
Li et al. (2017) [144] | Classification | PC, Intel E5-2650 2.0 GHz, 512 GB RAM, NVIDIA Tesla K20c (2496 CUDA cores). | CUDA (MAGMA), OpenBLAS (MPICH). | Various. COIL-100 (70,000 × 2500); MNIST (6,000,000 × 784). | 1.30; 0.35 | The performance of the heterogeneous CPU–GPU blocked algorithm is slightly better. |
Chen et al. (2017) [145] | Classification | Flink cluster, Intel Core i5-4590 3.30 GHz, 12 GB RAM, 2 × NVIDIA GeForce GTX 750 (512 CUDA cores). | Flink and CUDA. | Various. MNIST (6,000,000 × 784). | 0.35 | Performance on some GPUs is almost the same. |
Lam and Wunsch (2016) [146] | Classification | PC, Intel Xeon E5645 2.4 GHz, 12 GB RAM, NVIDIA Tesla M2075 (448 cores clocked) 1.5 GHz, 6 GB RAM. | CUDA. | CIFAR-10 (60,000 × 3072); MNIST (70,000 × 784). | 1.37; 0.41 | Each thread has a limited number of fast registers to store local variables. |
Reference | Application | Architecture | Parallel Tool | Data | Size (GB) * | Limitations |
---|---|---|---|---|---|---|
Wang et al. (2024) [97] | Classification | PC, Intel Core i7-10700 (8 core), 16 GB RAM. | MATLAB (2019). | Various. Gissete (7000 samples). | 0.26 | The algorithm computes and stores the matrix in each iteration, resulting in high computational cost and slow convergence. |
Wang and Soo (2023) [100] | Classification | PC, Intel(R) Xeon(R) Silver 4114 2.20 GHz, 16.0 GB RAM. | Python 3.8. | Various. Adults (44,222 samples). | 0.0046 | Optimal ELM number undefined, lacks online version, and adaptation to changing environments not developed. |
Zhang et al. (2023) [101] | Classification and regression | PC, AMD Ryzen 2600 3.40 GHz (6 cores), 16 GB RAM. | MATLAB (2021a). | Various. Mushroom (8124 samples). | 0.0013 | Further enhancements are needed, particularly in exploring additional regularization terms. |
Gelvez-Almeida et al. (2022–2023) [102,103] | Classification | Server, 2 × Intel(R) Xeon(R) Gold 2.20 GHz, 128 GB RAM. | OpenMP. | Synthetic fingerprint data set (2,000,000 samples). | 3.01 | It is necessary to compare with other ensemble models for validation and benchmarking. |
Zha et al. (2022) [107] | Regression | PC, Intel Core i5-9400F 2.9 GHz, 16 GB RAM. | MATLAB R2018b. | Various. SinC (5000 data). | 8.8 × 10−5 | The proposed model makes the model training time longer. |
Vidhya and Aji (2022) [108] | Classification | PC, Intel Core i5 2.90 GHz (6 core), 32 GB RAM. | Not reported | Various. RCV1 data set. | Up to 42.22 | The process of updating knowledge is not explored. |
Dong et al. (2021) [113] | Regression | PC, Intel Core i7-6700-K 3.4 GHz. | MATLAB and PSpice software platforms. | Various high-resolution images with 512 × 512 pixels. | 0.0003 | The passive-resistive network limits the size of the array and its use in memory design. |
Dwivedi et al. (2021) [151] | Classification | PC, Intel Core i7, 16 GB RAM. | MATLAB R2016. | NSL-KDD; AWID-ATK-R; NGIDS-DS data sets. | 0.052; 2.92; 5.04 | Only the basic architecture was used during the experiments. |
Zehai et al. (2021) [50] | Regression | Not reported. | Not reported. | Sinc and Mackey–Glass time-series data set. | Up to 0.002 | There are still many problems and challenges in other modular systems. |
Ezemobi et al. (2021) [114] | Regression | Texas F28379D microcontroller unit (MCU) board. | MATLAB. | 2 Ah capacity lithium-ion (does not report the number of samples). | Information not provided | The performance of the model is influenced by the choice of the discrete point-invariance interval of the input characteristics. |
Wu et al. (2021) [55] | Regression | PC, Intel Core i7-4700, 16 GB RAM. | R program. | Real data set (1966–2000 training and 2001–2015 testing). | Information not provided | The selection of the appropriate number of classifications remains a problem. |
Li et al. (2020) [116] | Regression | PC, 2.5 GHz CPU, 4 GB RAM. | MATLAB 2012a. | Various. Bank domains (8190 × 8); Elevator (8752 × 18); CBM (11,934 × 15). | 0.00024; 0.00058; 0.00066 | Decreasing learning is not considered in the model. |
Liang et al. (2019) [117] | Classification | Not reported. | Not reported. | It is obtained by the calibration experiment. | Information not provided | Excessive numbers of neurons will increase training time and may result in overfitting. |
Dokeroglu and Sevinc et al. (2019) [118] | Classification | Server, 64-bit CPU (8 cores), 256 GB RAM, 1.5 TB disk. | MPI. | Various. CHESS (3196 × 36); SPAM (4601 × 57). | 0.00043; 0.00097 | Future work proposes the use of hybrid metaheuristic algorithms supported by the GPU. |
Safaei et al. (2019) [48] | Classification | Xilinx Zynq platform. | System-on-a-chip (SoC) FPGA-based. | Housing (500 samples), Hollywood 3D and HON4D. | Information not provided | Floating-point is not used in this work. |
Li et al. (2018) [152] | Classification | PC, AMD Athlon TM X2 250 3.00 GHz, 2 GB RAM. | MATLAB 7.11.0 | Various. DNA (data with 180 features); Wine (data with 178 features). | Information not provided | Real-world applications should consider an online version of parallel ELM. |
Ming et al. (2018) [153] | Classification and regression | Supercomputer (2048 PCs), 2 × Intel Xeon X5670 2.93 GHz (6 cores), 48 GB RAM. | MPI and MKL (BLAS, PBLAS, ScaLAPACK, BLACS, among others.). | Various. YearPredictionMSD (515,345 × 90); MNIST8M (8.1M × 784). | 0.17; 23.65 | The method cannot support a large number of hidden neuron nodes, and the communication overhead is large. |
Henríquez and Ruz (2017) [154] | Classification and regression | PC, Intel Core i5 2.6 GHz, 8 GB RAM. | fOptions, RSNNS, MASS, and car. | Various. Skin (245,057 × 4); PPPT (45,730 × 9). | 0.0073; 0.003 | The performance of the proposed algorithm and ELM with some data sets is similar. |
Luo et al. (2017) [155] | Classification and regression | Server, 8 × 2.8 GHz, 8 GB RAM. | Not reported. | Various. Banknote (1372 × 4); Stock (950 × 12). | 0.000085; 0.00004 | More iterations with a larger number of processors are required to converge the objective function. |
Wang et al. (2016) [156] | Classification | Cluster (6 nodes), 2 × Intel Xeon E5-2640 2.5 GHz (12 cores), 64 GB RAM. | Armadillo, LAPACK, OpenBLAS, and MPICH. | Various. NORB (97,200 × 9216). | 3.34 | On the same data set, the upscaling is slightly reduced in the presence of more computational nodes. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gelvez-Almeida, E.; Mora, M.; Barrientos, R.J.; Hernández-García, R.; Vilches-Ponce, K.; Vera, M. A Review on Large-Scale Data Processing with Parallel and Distributed Randomized Extreme Learning Machine Neural Networks. Math. Comput. Appl. 2024, 29, 40. https://doi.org/10.3390/mca29030040
Gelvez-Almeida E, Mora M, Barrientos RJ, Hernández-García R, Vilches-Ponce K, Vera M. A Review on Large-Scale Data Processing with Parallel and Distributed Randomized Extreme Learning Machine Neural Networks. Mathematical and Computational Applications. 2024; 29(3):40. https://doi.org/10.3390/mca29030040
Chicago/Turabian StyleGelvez-Almeida, Elkin, Marco Mora, Ricardo J. Barrientos, Ruber Hernández-García, Karina Vilches-Ponce, and Miguel Vera. 2024. "A Review on Large-Scale Data Processing with Parallel and Distributed Randomized Extreme Learning Machine Neural Networks" Mathematical and Computational Applications 29, no. 3: 40. https://doi.org/10.3390/mca29030040
APA StyleGelvez-Almeida, E., Mora, M., Barrientos, R. J., Hernández-García, R., Vilches-Ponce, K., & Vera, M. (2024). A Review on Large-Scale Data Processing with Parallel and Distributed Randomized Extreme Learning Machine Neural Networks. Mathematical and Computational Applications, 29(3), 40. https://doi.org/10.3390/mca29030040