Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,109)

Search Parameters:
Keywords = gated recurrent unit

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 3360 KiB  
Article
Innovative Load Forecasting Models and Intelligent Control Strategy for Enhancing Distributed Load Levelling Techniques in Resilient Smart Grids
by Wang Fangzong and Zuhaib Nishtar
Electronics 2024, 13(17), 3552; https://doi.org/10.3390/electronics13173552 - 6 Sep 2024
Abstract
Dynamic load forecasting is essential for effective energy management and grid operation. The use of GRU (Gated Recurrent Unit) and Long Short-Term Memory (LSTM) networks for precise load prediction is investigated in this paper. This research examines dynamic load patterns by innovatively integrating [...] Read more.
Dynamic load forecasting is essential for effective energy management and grid operation. The use of GRU (Gated Recurrent Unit) and Long Short-Term Memory (LSTM) networks for precise load prediction is investigated in this paper. This research examines dynamic load patterns by innovatively integrating heterogeneous information from several datasets. The results show that the LSTM and GRU models are equally good at making predictions and that this holds true across a variety of datasets. Furthermore, the models’ ability to accurately capture the temporal relationships in the load data is demonstrated by their low Mean Absolute Percentage Error (MAPE) and Mean Squared Error (MSE) values. Additionally, the comparative analysis results, which highlight flexibility in model selection, can aid energy sector decision makers. The significance of precise load projections for maintaining grid dependability and optimizing resources is further highlighted by this work, which also elucidates the effects of forecast inaccuracies on decision-making procedures. Our research study provides important information for power system management strategy planning, which in turn promotes the continuous innovation of smart grids in dynamic load forecasting to keep up with changing energy consumption patterns. Full article
Show Figures

Figure 1

19 pages, 3854 KiB  
Article
Short-Term Campus Load Forecasting Using CNN-Based Encoder–Decoder Network with Attention
by Zain Ahmed, Mohsin Jamil and Ashraf Ali Khan
Energies 2024, 17(17), 4457; https://doi.org/10.3390/en17174457 - 5 Sep 2024
Viewed by 181
Abstract
Short-term load forecasting is a challenging research problem and has a tremendous impact on electricity generation, transmission, and distribution. A robust forecasting algorithm can help power system operators to better tackle the ever-changing electric power demand. This paper presents a novel deep neural [...] Read more.
Short-term load forecasting is a challenging research problem and has a tremendous impact on electricity generation, transmission, and distribution. A robust forecasting algorithm can help power system operators to better tackle the ever-changing electric power demand. This paper presents a novel deep neural network for short-term electric load forecasting for the St. John’s campus of Memorial University of Newfoundland (MUN). The electric load data are obtained from the Memorial University of Newfoundland and combined with metrological data from St. John’s. This dataset is used to formulate a multivariate time-series forecasting problem. A novel deep learning algorithm is presented, consisting of a 1D Convolutional Neural Network, which is followed by an encoder–decoder-based network with attention. The input used for this model is the electric load consumption and metrological data, while the output is the hourly prediction of the next day. The model is compared with Gated Recurrent Unit (GRU) and Long Short Term Memory (LSTM)-based Recurrent Neural Network. A CNN-based encoder–decoder model without attention is also tested. The proposed model shows a lower mean absolute error (MAE), mean squared error (MSE), mean absolute percentage error (MAPE), and higher R2 score. These evaluation metrics show an improved performance compared to GRU and LSTM-based RNNs as well as the CNN encoder–decoder model without attention. Full article
(This article belongs to the Special Issue Advances in Renewable Energy Power Forecasting and Integration)
Show Figures

Figure 1

20 pages, 8824 KiB  
Article
Research on Short-Term Forecasting Model of Global Atmospheric Temperature and Wind in the near Space Based on Deep Learning
by Xingxin Sun, Chen Zhou, Jian Feng, Huiyun Yang, Yuqiang Zhang, Zhou Chen, Tong Xu, Zhongxin Deng, Zhengyu Zhao, Yi Liu and Ting Lan
Atmosphere 2024, 15(9), 1069; https://doi.org/10.3390/atmos15091069 - 4 Sep 2024
Viewed by 171
Abstract
Developing short-term forecasting models for global atmospheric temperature and wind in near space is crucial for understanding atmospheric dynamics and supporting human activities in this region. While numerical models have been extensively developed, deep learning techniques have recently shown promise in improving atmospheric [...] Read more.
Developing short-term forecasting models for global atmospheric temperature and wind in near space is crucial for understanding atmospheric dynamics and supporting human activities in this region. While numerical models have been extensively developed, deep learning techniques have recently shown promise in improving atmospheric forecasting accuracy. In this study, convolutional long short-term memory (ConvLSTM) and convolutional gated recurrent unit (ConvGRU) neural networks were applied to build for short-term global-scale forecasting model of atmospheric temperature and wind in near space based on the MERRA-2 reanalysis dataset from 2010–2022. The model results showed that the ConvGRU model outperforms the ConvLSTM model in the short-term forecast results. The ConvGRU model achieved a root mean square error in the first three hours of approximately 1.8 K for temperature predictions, and errors of 4.2 m/s and 3.8 m/s for eastward and northward wind predictions on all 72 isobaric surfaces. Specifically, at a higher altitude (on the 1.65 Pa isobaric surface, approximately 70 km above sea level), the ConvGRU model achieved a RMSE of about 2.85 K for temperature predictions, and 5.67 m/s and 5.17 m/s for eastward and northward wind. This finding is significantly meaningful for short-term temperature and wind forecasts in near space and for exploring the physical mechanisms related to temperature and wind variations in this region. Full article
(This article belongs to the Section Atmospheric Techniques, Instruments, and Modeling)
Show Figures

Figure 1

29 pages, 31679 KiB  
Article
A Robust Recurrent Neural Networks-Based Surrogate Model for Thermal History and Melt Pool Characteristics in Directed Energy Deposition
by Sung-Heng Wu, Usman Tariq, Ranjit Joy, Muhammad Arif Mahmood, Asad Waqar Malik and Frank Liou
Materials 2024, 17(17), 4363; https://doi.org/10.3390/ma17174363 - 3 Sep 2024
Viewed by 466
Abstract
In directed energy deposition (DED), accurately controlling and predicting melt pool characteristics is essential for ensuring desired material qualities and geometric accuracies. This paper introduces a robust surrogate model based on recurrent neural network (RNN) architectures—Long Short-Term Memory (LSTM), Bidirectional LSTM (Bi-LSTM), and [...] Read more.
In directed energy deposition (DED), accurately controlling and predicting melt pool characteristics is essential for ensuring desired material qualities and geometric accuracies. This paper introduces a robust surrogate model based on recurrent neural network (RNN) architectures—Long Short-Term Memory (LSTM), Bidirectional LSTM (Bi-LSTM), and Gated Recurrent Unit (GRU). Leveraging a time series dataset from multi-physics simulations and a three-factor, three-level experimental design, the model accurately predicts melt pool peak temperatures, lengths, widths, and depths under varying conditions. RNN algorithms, particularly Bi-LSTM, demonstrate high predictive accuracy, with an R-square of 0.983 for melt pool peak temperatures. For melt pool geometry, the GRU-based model excels, achieving R-square values above 0.88 and reducing computation time by at least 29%, showcasing its accuracy and efficiency. The RNN-based surrogate model built in this research enhances understanding of melt pool dynamics and supports precise DED system setups. Full article
Show Figures

Figure 1

29 pages, 10253 KiB  
Article
Hyperspectral Image Denoising and Compression Using Optimized Bidirectional Gated Recurrent Unit
by Divya Mohan, Aravinth J and Sankaran Rajendran
Remote Sens. 2024, 16(17), 3258; https://doi.org/10.3390/rs16173258 - 2 Sep 2024
Viewed by 444
Abstract
The availability of a higher resolution fine spectral bandwidth in hyperspectral images (HSI) makes it easier to identify objects of interest in them. The inclusion of noise into the resulting collection of images is a limitation of HSI and has an adverse effect [...] Read more.
The availability of a higher resolution fine spectral bandwidth in hyperspectral images (HSI) makes it easier to identify objects of interest in them. The inclusion of noise into the resulting collection of images is a limitation of HSI and has an adverse effect on post-processing and data interpretation. Denoising HSI data is thus necessary for the effective execution of post-processing activities like image categorization and spectral unmixing. Most of the existing models cannot handle many forms of noise simultaneously. When it comes to compression, available compression models face the problems of increased processing time and lower accuracy. To overcome the existing limitations, an image denoising model using an adaptive fusion network is proposed. The denoised output is then processed through a compression model which uses an optimized deep learning technique called "chaotic Chebyshev artificial hummingbird optimization algorithm-based bidirectional gated recurrent unit" (CCAO-BiGRU). All the proposed models were tested in Python and evaluated using the Indian Pines, Washington DC Mall and CAVE datasets. The proposed model underwent qualitative and quantitative analysis and showed a PSNR value of 82 in the case of Indian Pines and 78.4 for the Washington DC Mall dataset at a compression rate of 10. The study proved that the proposed model provides the knowledge about complex nonlinear mapping between noise-free and noisy HSI for obtaining the denoised images and also results in high-quality compressed output. Full article
Show Figures

Figure 1

21 pages, 5219 KiB  
Article
Ensemble Learning for Nuclear Power Generation Forecasting Based on Deep Neural Networks and Support Vector Regression
by Jorge Gustavo Sandoval Simão and Leandro dos Santos Coelho
Technologies 2024, 12(9), 148; https://doi.org/10.3390/technologies12090148 - 2 Sep 2024
Viewed by 526
Abstract
Forecasting nuclear energy production is essential for market operations such as security, economic efficiency, resource optimization, grid stability, and the integration of renewable energy sources. Forecasting approaches allow nuclear power plants to operate consistently, contributing to the overall reliability and long-term viability of [...] Read more.
Forecasting nuclear energy production is essential for market operations such as security, economic efficiency, resource optimization, grid stability, and the integration of renewable energy sources. Forecasting approaches allow nuclear power plants to operate consistently, contributing to the overall reliability and long-term viability of the energy system. It is noted that energy systems researchers are increasingly interested in machine learning models used to face the challenge of time series forecasting. This study evaluates a hybrid ensemble learning of three time series forecasting models including least-squares support vector regression, gated recurrent unit, and long short-term memory models applied to nuclear power time series forecasting on the dataset of French power plants from 2009 to 2020. Furthermore, this research evaluates forecasting results in which approaches are directed towards the optimized RreliefF (Robust relief Feature) selection algorithm using a hyperparameter optimization based on tree-structured Parzen estimator and following an ensemble learning approach, showing promising results in terms of performance metrics. The suggested ensemble learning model, which combines deep learning and the RreliefF algorithm using a hold-out, outperforms the other nine forecasting models in this study according to performance criteria such as 75% for the coefficient of determination, a root squared error average of 0.108, and an average absolute error of 0.080. Full article
(This article belongs to the Collection Electrical Technologies)
Show Figures

Figure 1

24 pages, 5863 KiB  
Article
Utilizing TGAN and ConSinGAN for Improved Tool Wear Prediction: A Comparative Study with ED-LSTM, GRU, and CNN Models
by Milind Shah, Himanshu Borade, Vipul Dave, Hitesh Agrawal, Pranav Nair and Vinay Vakharia
Electronics 2024, 13(17), 3484; https://doi.org/10.3390/electronics13173484 - 2 Sep 2024
Viewed by 356
Abstract
Developing precise deep learning (DL) models for predicting tool wear is challenging, particularly due to the scarcity of experimental data. To address this issue, this paper introduces an innovative approach that leverages the capabilities of tabular generative adversarial networks (TGAN) and conditional single [...] Read more.
Developing precise deep learning (DL) models for predicting tool wear is challenging, particularly due to the scarcity of experimental data. To address this issue, this paper introduces an innovative approach that leverages the capabilities of tabular generative adversarial networks (TGAN) and conditional single image GAN (ConSinGAN). These models are employed to generate synthetic data, thereby enriching the dataset and enhancing the robustness of the predictive models. The efficacy of this methodology was rigorously evaluated using publicly available milling datasets. The pre-processing of acoustic emission data involved the application of the Walsh-Hadamard transform, followed by the generation of spectrograms. These spectrograms were then used to extract statistical attributes, forming a comprehensive feature vector for model input. Three DL models—encoder-decoder long short-term memory (ED-LSTM), gated recurrent unit (GRU), and convolutional neural network (CNN)—were applied to assess their tool wear prediction capabilities. The application of 10-fold cross-validation across these models yielded exceptionally low RMSE and MAE values of 0.02 and 0.16, respectively, underscoring the effectiveness of this approach. The results not only highlight the potential of TGAN and ConSinGAN in mitigating data scarcity but also demonstrate significant improvements in the accuracy of tool wear predictions, paving the way for more reliable and precise predictive maintenance in manufacturing processes. Full article
(This article belongs to the Special Issue New Advances in Machine Learning and Its Applications)
Show Figures

Figure 1

22 pages, 4175 KiB  
Article
Carbon Price Forecasting Using Optimized Sliding Window Empirical Wavelet Transform and Gated Recurrent Unit Network to Mitigate Data Leakage
by Zeyu Zhang, Xiaoqian Liu, Xiling Zhang, Zhishan Yang and Jian Yao
Energies 2024, 17(17), 4358; https://doi.org/10.3390/en17174358 - 31 Aug 2024
Viewed by 446
Abstract
Precise forecasts of carbon prices are crucial for reducing greenhouse gas emissions and promoting sustainable, low-carbon development. To mitigate noise interference in carbon price data, hybrid models integrating data decomposition techniques are commonly utilized. However, it has been observed that the improper utilization [...] Read more.
Precise forecasts of carbon prices are crucial for reducing greenhouse gas emissions and promoting sustainable, low-carbon development. To mitigate noise interference in carbon price data, hybrid models integrating data decomposition techniques are commonly utilized. However, it has been observed that the improper utilization of data decomposition techniques can lead to data leakage, thereby invalidating the model’s practical applicability. This study introduces a leakage-free hybrid model for carbon price forecasting based on the sliding window empirical wavelet transform (SWEWT) algorithm and the gated recurrent unit (GRU) network. First, the carbon price data are sampled using a sliding window approach and then decomposed into more stable and regular subcomponents through the EWT algorithm. By exclusively employing the data from the end of the window as input, the proposed method can effectively mitigate the risk of data leakage. Subsequently, the input data are passed into a multi-layer GRU model to extract patterns and features from the carbon price data. Finally, the optimized hybrid model is obtained by iteratively optimizing the hyperparameters of the model using the tree-structured Parzen estimator (TPE) algorithm, and the final prediction results are generated by the model. When used to forecast the closing price of the Guangdong Carbon Emission Allowance (GDEA) for the last nine years, the proposed hybrid model achieves outstanding performance with an R2 value of 0.969, significantly outperforming other structural variants. Furthermore, comparative experiments from various perspectives have validated the model’s structural rationality, practical applicability, and generalization capability, confirming that the proposed framework is a reliable choice for carbon price forecasting. Full article
(This article belongs to the Section B3: Carbon Emission and Utilization)
Show Figures

Figure 1

15 pages, 5646 KiB  
Article
Evaluation of Machine Learning Models in Air Pollution Prediction for a Case Study of Macau as an Effort to Comply with UN Sustainable Development Goals
by Thomas M. T. Lei, Jianxiu Cai, Altaf Hossain Molla, Tonni Agustiono Kurniawan and Steven Soon-Kai Kong
Sustainability 2024, 16(17), 7477; https://doi.org/10.3390/su16177477 - 29 Aug 2024
Viewed by 395
Abstract
To comply with the United Nations Sustainable Development Goals (UN SDGs), in particular with SDG 3, SDG 11, and SDG 13, a reliable air pollution prediction model must be developed to construct a sustainable, safe, and resilient city and mitigate climate change for [...] Read more.
To comply with the United Nations Sustainable Development Goals (UN SDGs), in particular with SDG 3, SDG 11, and SDG 13, a reliable air pollution prediction model must be developed to construct a sustainable, safe, and resilient city and mitigate climate change for a double win. Machine learning (ML) and deep learning (DL) models have been applied to datasets in Macau to predict the daily levels of roadside air pollution in the Macau peninsula, situated near the historical sites of Macau. Macau welcomed over 28 million tourists in 2023 as a popular tourism destination. Still, an accurate air quality forecast has not been in place for many years due to the lack of a reliable emission inventory. This work will develop a dependable air pollution prediction model for Macau, which is also the novelty of this study. The methods, including random forest (RF), support vector regression (SVR), artificial neural network (ANN), recurrent neural network (RNN), long short-term memory (LSTM), and gated recurrent unit (GRU), were applied and successful in the prediction of daily air pollution levels in Macau. The prediction model was trained using the air quality and meteorological data from 2013 to 2019 and validated using the data from 2020 to 2021. The model performance was evaluated based on the root mean square error (RMSE), mean absolute error (MAE), Pearson’s correlation coefficient (PCC), and Kendall’s tau coefficient (KTC). The RF model best predicted PM10, PM2.5, NO2, and CO concentrations with the highest PCC and KTC in a daily air pollution prediction. In addition, the SVR model had the best stability and repeatability compared to other models, with the lowest SD in RMSE, MAE, PCC, and KTC after five model runs. Therefore, the results of this study show that the RF model is more efficient and performs better than other models in the prediction of air pollution for the dataset of Macau. Full article
Show Figures

Figure 1

29 pages, 8131 KiB  
Article
Multi-Timeframe Forecasting Using Deep Learning Models for Solar Energy Efficiency in Smart Agriculture
by Saravanakumar Venkatesan and Yongyun Cho
Energies 2024, 17(17), 4322; https://doi.org/10.3390/en17174322 - 29 Aug 2024
Viewed by 385
Abstract
Since the advent of smart agriculture, technological advancements in solar energy have significantly improved farming practices, resulting in a substantial revival of different crop yields. However, the smart agriculture industry is currently facing challenges posed by climate change. This involves multi-timeframe forecasts for [...] Read more.
Since the advent of smart agriculture, technological advancements in solar energy have significantly improved farming practices, resulting in a substantial revival of different crop yields. However, the smart agriculture industry is currently facing challenges posed by climate change. This involves multi-timeframe forecasts for greenhouse operators covering short-, medium-, and long-term intervals. Solar energy not only reduces our reliance on non-renewable electricity but also plays a pivotal role in addressing climate change by lowering carbon emissions. This study aims to find a method to support consistently optimal solar energy use regardless of changes in greenhouse conditions by predicting solar energy (kWh) usage on various time steps. In this paper, we conducted solar energy usage prediction experiments on time steps using traditional Tensorflow Keras models (TF Keras), including a linear model (LM), Convolutional Neural Network (CNN), stacked—Long Short Term Memory (LSTM), stacked-Gated recurrent unit (GRU), and stacked-Bidirectional—Long Short —Term Memory (Bi-LSTM), as well as Tensor-Flow-based models for solar energy usage data from a smart farm. The stacked-Bi-LSTM outperformed the other DL models with Root Mean Squared Error (RMSE) of 0.0048, a Mean Absolute Error (MAE) of 0.0431, and R-Squared (R2) of 0.9243 in short-term prediction (2-h intervals). For mid-term (2-day) and long-term (2-week) forecasting, the stacked Bi-LSTM model also exhibited superior performance compared to other deep learning models, with RMSE values of 0.0257 and 0.0382, MAE values of 0.1103 and 0.1490, and R2 values of 0.5980 and 0.3974, respectively. The integration of multi-timeframe forecasting is expected to avoid conventional solar energy use forecasting, reduce the complexity of greenhouse energy management, and increase energy use efficiency compared to single-timeframe forecasting models. Full article
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)
Show Figures

Graphical abstract

7 pages, 2961 KiB  
Proceeding Paper
Preliminary Study on Image Captioning for Construction Hazards
by Wen-Ta Hsiao, Wen-Der Yu, Tao-Ming Cheng and Alexey Bulgakov
Eng. Proc. 2024, 74(1), 20; https://doi.org/10.3390/engproc2024074020 - 28 Aug 2024
Viewed by 77
Abstract
Construction accidents are a major contributor to occupational fatalities. To tackle this issue, improved monitoring for hazard elimination is crucial. By introducing a deep-learning image captioning system, we identified hazards via closed-circuit television in construction sites. By leveraging Inception-v3 for feature extraction and [...] Read more.
Construction accidents are a major contributor to occupational fatalities. To tackle this issue, improved monitoring for hazard elimination is crucial. By introducing a deep-learning image captioning system, we identified hazards via closed-circuit television in construction sites. By leveraging Inception-v3 for feature extraction and the gated recurrent unit for caption generation, real-time hazard monitoring was enabled. Bilingual evaluation understudy scores were determined for continuous and effective hazard detection, helping construction managers enhance safety. Full article
Show Figures

Figure 1

27 pages, 4478 KiB  
Article
Predicting Economic Trends and Stock Market Prices with Deep Learning and Advanced Machine Learning Techniques
by Victor Chang, Qianwen Ariel Xu, Anyamele Chidozie and Hai Wang
Electronics 2024, 13(17), 3396; https://doi.org/10.3390/electronics13173396 - 26 Aug 2024
Viewed by 1230
Abstract
The volatile and non-linear nature of stock market data, particularly in the post-pandemic era, poses significant challenges for accurate financial forecasting. To address these challenges, this research develops advanced deep learning and machine learning algorithms to predict financial trends, quantify risks, and forecast [...] Read more.
The volatile and non-linear nature of stock market data, particularly in the post-pandemic era, poses significant challenges for accurate financial forecasting. To address these challenges, this research develops advanced deep learning and machine learning algorithms to predict financial trends, quantify risks, and forecast stock prices, focusing on the technology sector. Our study seeks to answer the following question: “Which deep learning and supervised machine learning algorithms are the most accurate and efficient in predicting economic trends and stock market prices, and under what conditions do they perform best?” We focus on two advanced recurrent neural network (RNN) models, long short-term memory (LSTM) and Gated Recurrent Unit (GRU), to evaluate their efficiency in predicting technology industry stock prices. Additionally, we integrate statistical methods such as autoregressive integrated moving average (ARIMA) and Facebook Prophet and machine learning algorithms like Extreme Gradient Boosting (XGBoost) to enhance the robustness of our predictions. Unlike classical statistical algorithms, LSTM and GRU models can identify and retain important data sequences, enabling more accurate predictions. Our experimental results show that the GRU model outperforms the LSTM model in terms of prediction accuracy and training time across multiple metrics such as RMSE and MAE. This study offers crucial insights into the predictive capabilities of deep learning models and advanced machine learning techniques for financial forecasting, highlighting the potential of GRU and XGBoost for more accurate and efficient stock price prediction in the technology sector. Full article
Show Figures

Figure 1

34 pages, 786 KiB  
Review
Recurrent Neural Networks: A Comprehensive Review of Architectures, Variants, and Applications
by Ibomoiye Domor Mienye, Theo G. Swart and George Obaido
Information 2024, 15(9), 517; https://doi.org/10.3390/info15090517 - 25 Aug 2024
Viewed by 2018
Abstract
Recurrent neural networks (RNNs) have significantly advanced the field of machine learning (ML) by enabling the effective processing of sequential data. This paper provides a comprehensive review of RNNs and their applications, highlighting advancements in architectures, such as long short-term memory (LSTM) networks, [...] Read more.
Recurrent neural networks (RNNs) have significantly advanced the field of machine learning (ML) by enabling the effective processing of sequential data. This paper provides a comprehensive review of RNNs and their applications, highlighting advancements in architectures, such as long short-term memory (LSTM) networks, gated recurrent units (GRUs), bidirectional LSTM (BiLSTM), echo state networks (ESNs), peephole LSTM, and stacked LSTM. The study examines the application of RNNs to different domains, including natural language processing (NLP), speech recognition, time series forecasting, autonomous vehicles, and anomaly detection. Additionally, the study discusses recent innovations, such as the integration of attention mechanisms and the development of hybrid models that combine RNNs with convolutional neural networks (CNNs) and transformer architectures. This review aims to provide ML researchers and practitioners with a comprehensive overview of the current state and future directions of RNN research. Full article
(This article belongs to the Special Issue Applications of Machine Learning and Convolutional Neural Networks)
Show Figures

Figure 1

21 pages, 1042 KiB  
Article
Detection of Cyber-Attacks in a Discrete Event System Based on Deep Learning
by Sichen Ding, Gaiyun Liu, Li Yin, Jianzhou Wang and Zhiwu Li
Mathematics 2024, 12(17), 2635; https://doi.org/10.3390/math12172635 - 25 Aug 2024
Viewed by 444
Abstract
This paper addresses the problem of cyber-attack detection in a discrete event system by proposing a novel model. The model utilizes graph convolutional networks to extract spatial features from event sequences. Subsequently, it employs gated recurrent units to re-extract spatio-temporal features from these [...] Read more.
This paper addresses the problem of cyber-attack detection in a discrete event system by proposing a novel model. The model utilizes graph convolutional networks to extract spatial features from event sequences. Subsequently, it employs gated recurrent units to re-extract spatio-temporal features from these spatial features. The obtained spatio-temporal features are then fed into an attention model. This approach enables the model to learn the importance of different event sequences, ensuring that it is sufficiently general for identifying cyber-attacks, obviating the need to specify attack types. Compared with traditional methods that rely on synchronous product computations to synthesize diagnosers, our deep learning-based model circumvents state explosion problems. Our method facilitates real-time and efficient cyber-attack detection, eliminating the necessity to specifically identify system states or distinguish attack types, thereby significantly simplifying the diagnostic process. Additionally, we set an adjustable probability threshold to determine whether an event sequence has been compromised, allowing for customization to meet diverse requirements. Experimental results demonstrate that the proposed method performs well in cyber-attack detection, achieving over 99.9% accuracy at a 1% threshold and a weighted F1-score of 0.8126, validating its superior performance. Full article
(This article belongs to the Special Issue Discrete Event Dynamic Systems and Applications)
Show Figures

Figure 1

24 pages, 21090 KiB  
Article
Comparative Analysis of Machine-Learning Models for Soil Moisture Estimation Using High-Resolution Remote-Sensing Data
by Ming Li and Yueguan Yan
Land 2024, 13(8), 1331; https://doi.org/10.3390/land13081331 - 22 Aug 2024
Viewed by 511
Abstract
Soil moisture is an important component of the hydrologic cycle and ecosystem functioning, and it has a significant impact on agricultural production, climate change and natural disasters. Despite the availability of machine-learning techniques for estimating soil moisture from high-resolution remote-sensing imagery, including synthetic [...] Read more.
Soil moisture is an important component of the hydrologic cycle and ecosystem functioning, and it has a significant impact on agricultural production, climate change and natural disasters. Despite the availability of machine-learning techniques for estimating soil moisture from high-resolution remote-sensing imagery, including synthetic aperture radar (SAR) data and optical remote sensing, comprehensive comparative studies of these techniques remain limited. This paper addresses this gap by systematically comparing the performance of four tree-based ensemble-learning models (random forest (RF), extreme gradient boosting (XGBoost), light gradient-boosting machine (LightGBM), and category boosting (CatBoost)) and three deep-learning models (deep neural network (DNN), convolutional neural network (CNN), and gated recurrent unit (GRU)) in terms of soil moisture estimation. Additionally, we introduce and evaluate the effectiveness of four different stacking methods for model fusion, an approach that is relatively novel in this context. Moreover, Sentinel-1 C-band dual-polarization SAR and Sentinel-2 multispectral data, as well as NASADEM and geographical code and temporal code features, are used as input variables to retrieve the soil moisture in the ShanDian River Basin in China. Our findings reveal that the tree-based ensemble-learning models outperform the deep-learning models, with LightGBM being the best individual model, while the stacking approach can further enhance the accuracy and robustness of soil moisture estimation. Moreover, the stacking all boosting classes ensemble-learning model (SABM), which integrates only boosting-type models, demonstrates superior accuracy and robustness in soil moisture estimation. The SHAP value analysis reveals that ensemble learning can utilize more complex features than deep learning. This study provides an effective method for retrieving soil moisture using machine-learning and high-resolution remote-sensing data, demonstrating the application value of SAR data and high-resolution optical remote-sensing data in soil moisture monitoring. Full article
Show Figures

Figure 1

Back to TopTop