Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,148)

Search Parameters:
Keywords = Bi-LSTM

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 10557 KiB  
Article
Identification of Anomalies in Lung and Colon Cancer Using Computer Vision-Based Swin Transformer with Ensemble Model on Histopathological Images
by Abdulkream A. Alsulami, Aishah Albarakati, Abdullah AL-Malaise AL-Ghamdi and Mahmoud Ragab
Bioengineering 2024, 11(10), 978; https://doi.org/10.3390/bioengineering11100978 (registering DOI) - 28 Sep 2024
Abstract
Lung and colon cancer (LCC) is a dominant life-threatening disease that needs timely attention and precise diagnosis for efficient treatment. The conventional diagnostic techniques for LCC regularly encounter constraints in terms of efficiency and accuracy, thus causing challenges in primary recognition and treatment. [...] Read more.
Lung and colon cancer (LCC) is a dominant life-threatening disease that needs timely attention and precise diagnosis for efficient treatment. The conventional diagnostic techniques for LCC regularly encounter constraints in terms of efficiency and accuracy, thus causing challenges in primary recognition and treatment. Early diagnosis of the disease can immensely reduce the probability of death. In medical practice, the histopathological study of the tissue samples generally uses a classical model. Still, the automated devices that exploit artificial intelligence (AI) techniques produce efficient results in disease diagnosis. In histopathology, both machine learning (ML) and deep learning (DL) approaches can be deployed owing to their latent ability in analyzing and predicting physically accurate molecular phenotypes and microsatellite uncertainty. In this background, this study presents a novel technique called Lung and Colon Cancer using a Swin Transformer with an Ensemble Model on the Histopathological Images (LCCST-EMHI). The proposed LCCST-EMHI method focuses on designing a DL model for the diagnosis and classification of the LCC using histopathological images (HI). In order to achieve this, the LCCST-EMHI model utilizes the bilateral filtering (BF) technique to get rid of the noise. Further, the Swin Transformer (ST) model is also employed for the purpose of feature extraction. For the LCC detection and classification process, an ensemble deep learning classifier is used with three techniques: bidirectional long short-term memory with multi-head attention (BiLSTM-MHA), Double Deep Q-Network (DDQN), and sparse stacked autoencoder (SSAE). Eventually, the hyperparameter selection of the three DL models can be implemented utilizing the walrus optimization algorithm (WaOA) method. In order to illustrate the promising performance of the LCCST-EMHI approach, an extensive range of simulation analyses was conducted on a benchmark dataset. The experimentation results demonstrated the promising performance of the LCCST-EMHI approach over other recent methods. Full article
(This article belongs to the Special Issue Computer Vision and Machine Learning in Medical Applications)
Show Figures

Figure 1

15 pages, 1823 KiB  
Article
Enhancing Recommendation Diversity and Novelty with Bi-LSTM and Mean Shift Clustering
by Yuan Yuan, Yuying Zhou, Xuanyou Chen, Qi Xiong and Hector Chimeremeze Okere
Electronics 2024, 13(19), 3841; https://doi.org/10.3390/electronics13193841 (registering DOI) - 28 Sep 2024
Abstract
In the digital age, personalized recommendation systems have become crucial for information dissemination and user experience. While traditional systems focus on accuracy, they often overlook diversity, novelty, and serendipity. This study introduces an innovative recommendation system model, Time-based Outlier Aware Recommender (TOAR), designed [...] Read more.
In the digital age, personalized recommendation systems have become crucial for information dissemination and user experience. While traditional systems focus on accuracy, they often overlook diversity, novelty, and serendipity. This study introduces an innovative recommendation system model, Time-based Outlier Aware Recommender (TOAR), designed to address the challenges of content homogenization and information bubbles in personalized recommendations. TOAR integrates Neural Matrix Factorization (NeuMF), Bidirectional Long Short-Term Memory Networks (Bi-LSTM), and Mean Shift clustering to enhance recommendation accuracy, novelty, and diversity. The model analyzes temporal dynamics of user behavior and facilitates cross-domain knowledge exchange through feature sharing and transfer learning mechanisms. By incorporating an attention mechanism and unsupervised clustering, TOAR effectively captures important time-series information and ensures recommendation diversity. Experimental results on a news recommendation dataset demonstrate TOAR’s superior performance across multiple metrics, including AUC, precision, NDCG, and novelty, compared to traditional and deep learning-based recommendation models. This research provides a foundation for developing more intelligent and personalized recommendation services that balance accuracy with content diversity. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

13 pages, 4569 KiB  
Article
End-to-End Electrocardiogram Signal Transformation from Continuous-Wave Radar Signal Using Deep Learning Model with Maximum-Overlap Discrete Wavelet Transform and Adaptive Neuro-Fuzzy Network Layers
by Tae-Wan Kim and Keun-Chang Kwak
Appl. Sci. 2024, 14(19), 8730; https://doi.org/10.3390/app14198730 - 27 Sep 2024
Viewed by 189
Abstract
This paper is concerned with an end-to-end electrocardiogram (ECG) signal transformation from a continuous-wave (CW) radar signal using a specialized deep learning model. For this purpose, the presented deep learning model is designed using convolutional neural networks (CNNs) and bidirectional long short-term memory [...] Read more.
This paper is concerned with an end-to-end electrocardiogram (ECG) signal transformation from a continuous-wave (CW) radar signal using a specialized deep learning model. For this purpose, the presented deep learning model is designed using convolutional neural networks (CNNs) and bidirectional long short-term memory (Bi-LSTM) with a maximum-overlap discrete wavelet transform (MODWT) layer and an adaptive neuro-fuzzy network (ANFN) layer. The proposed method has the advantage of developing existing deep networks and machine learning to reconstruct signals through CW radars to acquire ECG biological information in a non-contact manner. The fully connected (FC) layer of the CNN is replaced by an ANFN layer suitable for resolving black boxes and handling complex nonlinear data. The MODWT layer is activated via discrete wavelet transform frequency decomposition with maximum-overlap to extract ECG-related frequency components from radar signals to generate essential information. In order to evaluate the performance of the proposed model, we use a dataset of clinically recorded vital signs with a synchronized reference sensor signal measured simultaneously. As a result of the experiment, the performance is evaluated by the mean squared error (MSE) between the measured and reconstructed ECG signals. The experimental results reveal that the proposed model shows good performance in comparison to the existing deep learning model. From the performance comparison, we confirm that the ANFN layer preserves the nonlinearity of information received from the model by replacing the fully connected layer used in the conventional deep learning model. Full article
Show Figures

Figure 1

27 pages, 3914 KiB  
Article
Temporally Correlated Deep Learning-Based Horizontal Wind-Speed Prediction
by Lintong Li, Jose Escribano-Macias, Mingwei Zhang, Shenghao Fu, Mingyang Huang, Xiangmin Yang, Tianyu Zhao, Yuxiang Feng, Mireille Elhajj, Arnab Majumdar, Panagiotis Angeloudis and Washington Ochieng
Sensors 2024, 24(19), 6254; https://doi.org/10.3390/s24196254 - 27 Sep 2024
Viewed by 265
Abstract
Wind speed affects aviation performance, clean energy production, and other applications. By accurately predicting wind speed, operational delays and accidents can be avoided, while the efficiency of wind energy production can also be increased. This paper initially overviews the definition, characteristics, sensors capable [...] Read more.
Wind speed affects aviation performance, clean energy production, and other applications. By accurately predicting wind speed, operational delays and accidents can be avoided, while the efficiency of wind energy production can also be increased. This paper initially overviews the definition, characteristics, sensors capable of measuring the feature, and the relationship between this feature and wind speed for all Quality Indicators (QIs). Subsequently, the feature importance of each QI relevant to wind-speed prediction is assessed, and all QIs are employed to predict horizontal wind speed. In addition, we conduct a comparison between the performance of traditional point-wise machine learning models and temporally correlated deep learning ones. The results demonstrate that the Bidirectional Long Short-Term Memory (BiLSTM) neural network yielded the highest level of accuracy across three metrics. Additionally, the newly proposed set of QIs outperformed the previously utilised QIs to a significant degree. Full article
(This article belongs to the Section Physical Sensors)
Show Figures

Figure 1

20 pages, 22656 KiB  
Article
Intelligent Reflecting Surface-Assisted Wireless Communication Using RNNs: Comprehensive Insights
by Rana Tabassum, Mohammad Abrar Shakil Sejan, Md Habibur Rahman, Md Abdul Aziz and Hyoung-Kyu Song
Mathematics 2024, 12(19), 2973; https://doi.org/10.3390/math12192973 - 25 Sep 2024
Viewed by 320
Abstract
By adjusting the propagation environment using reconfigurable reflecting elements, intelligent reflecting surfaces (IRSs) have become potential techniques used to improve the efficiency of wireless communication networks. In IRS-assisted communication systems, accurate channel estimation is crucial for optimizing signal transmission and achieving high spectral [...] Read more.
By adjusting the propagation environment using reconfigurable reflecting elements, intelligent reflecting surfaces (IRSs) have become potential techniques used to improve the efficiency of wireless communication networks. In IRS-assisted communication systems, accurate channel estimation is crucial for optimizing signal transmission and achieving high spectral efficiency. As mobile data traffic continues to surge and the demand for high-capacity and low-latency wireless connectivity grows, IRSs are becoming pivotal technologies in the development of next-generation communication networks. IRSs offer the potential to revolutionize wireless propagation environments, improving network capacity and coverage, particularly in high-frequency wave scenarios where traditional signals encounter obstacles. Amidst this evolving landscape, machine learning (ML) emerges as a powerful tool to harness the full potential of IRS-assisted communication systems, particularly given the escalating computational complexity associated with deploying and operating IRSs in dynamic environments. This paper presents an overview of preliminary results for IRS-assisted communication using recurrent neural networks (RNNs). We first implement single- and double-layer LSTM, BiLSTM, and GRU techniques for an IRS-based communication system. In the next phase, we explore a hybrid approach, combining different RNN techniques, including LSTM-BiLSTM, LSTM-GRU, and BiLSTM-GRU, as well as their reverse configurations. These RNN algorithms were evaluated with respect to bit error rate (BER) and symbol error rate (SER) for IRS-enhanced communication. According to the experimental results, the BiLSTM double-layer model and the BiLSTM-GRU combination demonstrated the highest BER and SER accuracy compared to other approaches. Full article
Show Figures

Figure 1

13 pages, 2319 KiB  
Article
Fault Diagnosis for Motor Bearings via an Intelligent Strategy Combined with Signal Reconstruction and Deep Learning
by Weiguo Li, Naiyuan Fan, Xiang Peng, Changhong Zhang, Mingyang Li, Xu Yang and Lijuan Ma
Energies 2024, 17(19), 4773; https://doi.org/10.3390/en17194773 - 24 Sep 2024
Viewed by 283
Abstract
To overcome the incomplete decomposition of vibration signals in traditional motor-bearing fault diagnosis algorithms and improve the ability to characterize fault characteristics and anti-interference, a diagnostic strategy combining dual signal reconstruction and deep learning architecture is proposed. In this study, an improved complete [...] Read more.
To overcome the incomplete decomposition of vibration signals in traditional motor-bearing fault diagnosis algorithms and improve the ability to characterize fault characteristics and anti-interference, a diagnostic strategy combining dual signal reconstruction and deep learning architecture is proposed. In this study, an improved complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) and variational mode decomposition (VMD)-based signal reconstruction method is first introduced to extract features representing motor bearing faults. A feature matrix construction method based on improved information entropy is then proposed to quantify these fault features. Finally, a fault diagnosis algorithm architecture integrating a multi-scale convolutional neural network (MSCNN) with attention mechanisms and a bidirectional long short-term memory network (BiLSTM) is developed. The experimental results for four fault states show that this model can effectively extract fault features from original vibration signals and, compared to other fault diagnosis models, offer high diagnostic accuracy and strong generalization, maintaining high accuracy even under varying speeds and noise interference. Full article
21 pages, 1256 KiB  
Article
Predicting Power Consumption Using Deep Learning with Stationary Wavelet
by Majdi Frikha, Khaled Taouil, Ahmed Fakhfakh and Faouzi Derbel
Forecasting 2024, 6(3), 864-884; https://doi.org/10.3390/forecast6030043 - 23 Sep 2024
Viewed by 462
Abstract
Power consumption in the home has grown in recent years as a consequence of the use of varied residential applications. On the other hand, many families are beginning to use renewable energy, such as energy production, energy storage devices, and electric vehicles. As [...] Read more.
Power consumption in the home has grown in recent years as a consequence of the use of varied residential applications. On the other hand, many families are beginning to use renewable energy, such as energy production, energy storage devices, and electric vehicles. As a result, estimating household power demand is necessary for energy consumption monitoring and planning. Power consumption forecasting is a challenging time series prediction topic. Furthermore, conventional forecasting approaches make it difficult to anticipate electric power consumption since it comprises irregular trend components, such as regular seasonal fluctuations. To address this issue, algorithms combining stationary wavelet transform (SWT) with deep learning models have been proposed. The denoised series is fitted with various benchmark models, including Long Short-Term Memory (LSTM), Gated Recurrent Units (GRUs), Bidirectional Gated Recurrent Units (Bi-GRUs), Bidirectional Long Short-Term Memory (Bi-LSTM), and Bidirectional Gated Recurrent Units Long Short-Term Memory (Bi-GRU LSTM) models. The performance of the SWT approach is evaluated using power consumption data at three different time intervals (1 min, 15 min, and 1 h). The performance of these models is evaluated using metrics such as Root Mean Square Error (RMSE), Mean Absolute Percentage Error (MAPE), and Mean Absolute Error (MAE). The SWT/GRU model, utilizing the bior2.4 filter at level 1, has emerged as a highly reliable option for precise power consumption forecasting across various time intervals. It is observed that the bior2.4/GRU model has enhanced accuracy by over 60% compared to the deep learning model alone across all accuracy measures. The findings clearly highlight the success of the SWT denoising technique with the bior2.4 filter in improving the power consumption prediction accuracy. Full article
(This article belongs to the Section Power and Energy Forecasting)
Show Figures

Figure 1

30 pages, 8653 KiB  
Article
CGAOA-AttBiGRU: A Novel Deep Learning Framework for Forecasting CO2 Emissions
by Haijun Liu, Yang Wu, Dongqing Tan, Yi Chen and Haoran Wang
Mathematics 2024, 12(18), 2956; https://doi.org/10.3390/math12182956 - 23 Sep 2024
Viewed by 268
Abstract
Accurately predicting carbon dioxide (CO2) emissions is crucial for environmental protection. Currently, there are two main issues with predicting CO2 emissions: (1) existing CO2 emission prediction models mainly rely on Long Short-Term Memory (LSTM) and Gate Recurrent Unit (GRU) [...] Read more.
Accurately predicting carbon dioxide (CO2) emissions is crucial for environmental protection. Currently, there are two main issues with predicting CO2 emissions: (1) existing CO2 emission prediction models mainly rely on Long Short-Term Memory (LSTM) and Gate Recurrent Unit (GRU) models, which can only model unidirectional temporal features, resulting in insufficient accuracy: (2) existing research on CO2 emissions mainly focuses on designing predictive models, without paying attention to model optimization, resulting in models being unable to achieve their optimal performance. To address these issues, this paper proposes a framework for predicting CO2 emissions, called CGAOA-AttBiGRU. In this framework, Attentional-Bidirectional Gate Recurrent Unit (AttBiGRU) is a prediction model that uses BiGRU units to extract bidirectional temporal features from the data, and adopts an attention mechanism to adaptively weight the bidirectional temporal features, thereby improving prediction accuracy. CGAOA is an improved Arithmetic Optimization Algorithm (AOA) used to optimize the five key hyperparameters of the AttBiGRU. We first validated the optimization performance of the improved CGAOA algorithm on 24 benchmark functions. Then, CGAOA was used to optimize AttBiGRU and compared with 12 optimization algorithms. The results indicate that the AttBiGRU optimized by CGAOA has the best predictive performance. Full article
Show Figures

Figure 1

19 pages, 15139 KiB  
Article
Ultra-Short-Term Photovoltaic Power Prediction by NRGA-BiLSTM Considering Seasonality and Periodicity of Data
by Hong Wu, Haipeng Liu, Huaiping Jin and Yanping He
Energies 2024, 17(18), 4739; https://doi.org/10.3390/en17184739 - 23 Sep 2024
Viewed by 523
Abstract
Photovoltaic (PV) power generation is highly stochastic and intermittent, which poses a challenge to the planning and operation of existing power systems. To enhance the accuracy of PV power prediction and ensure the safe operation of the power system, a novel approach based [...] Read more.
Photovoltaic (PV) power generation is highly stochastic and intermittent, which poses a challenge to the planning and operation of existing power systems. To enhance the accuracy of PV power prediction and ensure the safe operation of the power system, a novel approach based on seasonal division and a periodic attention mechanism (PAM) for PV power prediction is proposed. First, the dataset is divided into three components of trend, period, and residual under fuzzy c-means clustering (FCM) and the seasonal decomposition (SD) method according to four seasons. Three independent bidirectional long short-term memory (BiLTSM) networks are constructed for these subsequences. Then, the network is optimized using the improved Newton–Raphson genetic algorithm (NRGA), and the innovative PAM is added to focus on the periodic characteristics of the data. Finally, the results of each component are summarized to obtain the final prediction results. A case study of the Australian DKASC Alice Spring PV power plant dataset demonstrates the performance of the proposed approach. Compared with other paper models, the MAE, RMSE, and MAPE performance evaluation indexes show that the proposed approach has excellent performance in predicting output power accuracy and stability. Full article
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)
Show Figures

Figure 1

37 pages, 11393 KiB  
Article
Optimizing Deep Learning Models with Improved BWO for TEC Prediction
by Yi Chen, Haijun Liu, Weifeng Shan, Yuan Yao, Lili Xing, Haoran Wang and Kunpeng Zhang
Biomimetics 2024, 9(9), 575; https://doi.org/10.3390/biomimetics9090575 - 22 Sep 2024
Viewed by 414
Abstract
The prediction of total ionospheric electron content (TEC) is of great significance for space weather monitoring and wireless communication. Recently, deep learning models have become increasingly popular in TEC prediction. However, these deep learning models usually contain a large number of hyperparameters. Finding [...] Read more.
The prediction of total ionospheric electron content (TEC) is of great significance for space weather monitoring and wireless communication. Recently, deep learning models have become increasingly popular in TEC prediction. However, these deep learning models usually contain a large number of hyperparameters. Finding the optimal hyperparameters (also known as hyperparameter optimization) is currently a great challenge, directly affecting the predictive performance of the deep learning models. The Beluga Whale Optimization (BWO) algorithm is a swarm intelligence optimization algorithm that can be used to optimize hyperparameters of deep learning models. However, it is easy to fall into local minima. This paper analyzed the drawbacks of BWO and proposed an improved BWO algorithm, named FAMBWO (Firefly Assisted Multi-strategy Beluga Whale Optimization). Our proposed FAMBWO was compared with 11 state-of-the-art swarm intelligence optimization algorithms on 30 benchmark functions, and the results showed that our improved algorithm had faster convergence speed and better solutions on almost all benchmark functions. Then we proposed an automated machine learning framework FAMBWO-MA-BiLSTM for TEC prediction, where MA-BiLSTM is for TEC prediction and FAMBWO for hyperparameters optimization. We compared it with grid search, random search, Bayesian optimization algorithm and beluga whale optimization algorithm. Results showed that the MA-BiLSTM model optimized by FAMBWO is significantly better than the MA-BiLSTM model optimized by grid search, random search, Bayesian optimization algorithm, and BWO. Full article
Show Figures

Figure 1

16 pages, 5554 KiB  
Article
Improving the ERA5-Land Temperature Product through a Deep Spatiotemporal Model That Uses Fused Multi-Source Remote Sensing Data
by Lei Xu, Jinjin Du, Jiwei Ren, Qiannan Hu, Fen Qin, Weichen Mu and Jiyuan Hu
Remote Sens. 2024, 16(18), 3510; https://doi.org/10.3390/rs16183510 - 21 Sep 2024
Viewed by 593
Abstract
Temperature is a crucial indicator for studying climate, as well as the social and economic changes in a region. Temperature reanalysis products, such as ERA5-Land, have been widely used in studying temperature change. However, global-scale temperature reanalysis products have errors because they overlook [...] Read more.
Temperature is a crucial indicator for studying climate, as well as the social and economic changes in a region. Temperature reanalysis products, such as ERA5-Land, have been widely used in studying temperature change. However, global-scale temperature reanalysis products have errors because they overlook the influence of multiple factors on temperature, and this issue is more obvious in smaller areas. During the cold months (January, February, March, November, and December) in the Yellow River Basin, ERA5-Land products exhibit significant errors compared to temperatures observed by meteorological stations, typically underestimating the temperature. This study proposes improving temperature reanalysis products using deep learning and multi-source remote sensing and geographic data fusion. Specifically, convolutional neural networks (CNN) and bidirectional long short-term memory networks (BiLSTM) capture the spatial and temporal relationships between temperature, DEM, land cover, and population density. A deep spatiotemporal model is established to enhance temperature reanalysis products, resulting in higher resolution and more accurate temperature data. A comparison with the measured temperatures at meteorological stations indicates that the accuracy of the improved ERA5-Land product has been significantly enhanced, with the mean absolute error (MAE) reduced by 28.7% and the root mean square error (RMSE) reduced by 25.8%. This method obtained a high-precision daily temperature dataset with a 0.05° resolution for cold months in the Yellow River Basin from 2015 to 2019. Based on this dataset, the annual trend of average temperature changes during the cold months in the Yellow River Basin was analyzed. This study provides a scientific basis for improving ERA5-Land temperature reanalysis products in the Yellow River Basin and offers theoretical support for climate change research in the region. Full article
(This article belongs to the Special Issue Deep Learning for Remote Sensing and Geodata)
Show Figures

Figure 1

23 pages, 3894 KiB  
Article
Real-Time Fire Classification Models Based on Deep Learning for Building an Intelligent Multi-Sensor System
by Youngchan Kim, Yoseob Heo, Byoungsam Jin and Youngchul Bae
Fire 2024, 7(9), 329; https://doi.org/10.3390/fire7090329 - 21 Sep 2024
Viewed by 374
Abstract
Fire detection systems are critical for mitigating the damage caused by fires, which can result in significant annual property losses and fatalities. This paper presents a deep learning-based fire classification model for an intelligent multi-sensor system aimed at early and reliable fire detection. [...] Read more.
Fire detection systems are critical for mitigating the damage caused by fires, which can result in significant annual property losses and fatalities. This paper presents a deep learning-based fire classification model for an intelligent multi-sensor system aimed at early and reliable fire detection. The model processes data from multiple sensors that detect various parameters, such as temperature, humidity, and gas concentrations. Several deep learning architectures were evaluated, including LSTM, GRU, Bi-LSTM, LSTM-FCN, InceptionTime, and Transformer. The models were trained on data collected from controlled fire scenarios and validated for classification accuracy, loss, and real-time performance. The results indicated that the LSTM-based models (particularly Bi-LSTM and LSTM) could achieve high classification accuracy and low false alarm rates, demonstrating their effectiveness for real-time fire detection. The findings highlight the potential of advanced deep-learning models to enhance the reliability of sensor-based fire detection systems. Full article
(This article belongs to the Special Issue Advances in Building Fire Safety Engineering)
Show Figures

Figure 1

23 pages, 7374 KiB  
Article
A Chinese Nested Named Entity Recognition Model for Chicken Disease Based on Multiple Fine-Grained Feature Fusion and Efficient Global Pointer
by Xiajun Wang, Cheng Peng, Qifeng Li, Qinyang Yu, Liqun Lin, Pingping Li, Ronghua Gao, Wenbiao Wu, Ruixiang Jiang, Ligen Yu, Luyu Ding and Lei Zhu
Appl. Sci. 2024, 14(18), 8495; https://doi.org/10.3390/app14188495 - 20 Sep 2024
Viewed by 520
Abstract
Extracting entities from large volumes of chicken epidemic texts is crucial for knowledge sharing, integration, and application. However, named entity recognition (NER) encounters significant challenges in this domain, particularly due to the prevalence of nested entities and domain-specific named entities, coupled with a [...] Read more.
Extracting entities from large volumes of chicken epidemic texts is crucial for knowledge sharing, integration, and application. However, named entity recognition (NER) encounters significant challenges in this domain, particularly due to the prevalence of nested entities and domain-specific named entities, coupled with a scarcity of labeled data. To address these challenges, we compiled a corpus from 50 books on chicken diseases, covering 28 different disease types. Utilizing this corpus, we constructed the CDNER dataset and developed a nested NER model, MFGFF-BiLSTM-EGP. This model integrates the multiple fine-grained feature fusion (MFGFF) module with a BiLSTM neural network and employs an efficient global pointer (EGP) to predict the entity location encoding. In the MFGFF module, we designed three encoders: the character encoder, word encoder, and sentence encoder. This design effectively captured fine-grained features and improved the recognition accuracy of nested entities. Experimental results showed that the model performed robustly, with F1 scores of 91.98%, 73.32%, and 82.54% on the CDNER, CMeEE V2, and CLUENER datasets, respectively, outperforming other commonly used NER models. Specifically, on the CDNER dataset, the model achieved an F1 score of 79.68% for nested entity recognition. This research not only advances the development of a knowledge graph and intelligent question-answering system for chicken diseases, but also provides a viable solution for extracting disease information that can be applied to other livestock species. Full article
(This article belongs to the Special Issue Applied Intelligence in Natural Language Processing)
Show Figures

Figure 1

15 pages, 5583 KiB  
Article
The Development of Bi-LSTM Based on Fault Diagnosis Scheme in MVDC System
by Jae-Sung Lim, Haesong Cho, Dohoon Kwon and Junho Hong
Energies 2024, 17(18), 4689; https://doi.org/10.3390/en17184689 - 20 Sep 2024
Viewed by 425
Abstract
Diagnosing faults is crucial for ensuring the safety and reliability of medium-voltage direct current (MVDC) systems. In this study, we propose a bidirectional long short-term memory (Bi-LSTM)-based fault diagnosis scheme for the accurate classification of faults occurring in MVDC systems. First, to ensure [...] Read more.
Diagnosing faults is crucial for ensuring the safety and reliability of medium-voltage direct current (MVDC) systems. In this study, we propose a bidirectional long short-term memory (Bi-LSTM)-based fault diagnosis scheme for the accurate classification of faults occurring in MVDC systems. First, to ensure stability in case a fault occurs, we modeled an MVDC system that included a resistor-based fault current limiter (R-FCL) and a direct current circuit breaker (DCCB). A discrete wavelet transform (DWT) extracted the transient voltages and currents measured using DC lines and AC grids in the frequency–time domain. Based on the digital signal normalized by the DWT, using the measurement data, the Bi-LSTM algorithm was used to classify and learn the types and locations of faults, such as DC line (PTP, P-PTG, and N-PTG) and internal inverter faults. The effectiveness of the proposed fault diagnosis scheme was validated through comparative analysis within the four-terminal MVDC system, demonstrating superior accuracy and a faster diagnosis time compared to those of the existing schemes that utilize other AI algorithms, such as the CNN and LSTM. According to the test results, the proposed fault diagnosis scheme detects MVDC faults and shows a high recognition accuracy of 97.7%. Additionally, when applying the Bi-LSTM-based fault diagnosis scheme, it was confirmed that not only the training diagnosis time (TraDT) but also the average diagnosis time (AvgDT) were 0.03 ms and 0.05 ms faster than LSTM and CNN, respectively. The results validate the superior fault clarification and fast diagnosis performance of the proposed scheme over those of the other methods. Full article
(This article belongs to the Special Issue Advances in Research and Practice of Smart Electric Power Systems)
Show Figures

Figure 1

19 pages, 644 KiB  
Article
SMS Scam Detection Application Based on Optical Character Recognition for Image Data Using Unsupervised and Deep Semi-Supervised Learning
by Anjali Shinde, Essa Q. Shahra, Shadi Basurra, Faisal Saeed, Abdulrahman A. AlSewari and Waheb A. Jabbar
Sensors 2024, 24(18), 6084; https://doi.org/10.3390/s24186084 - 20 Sep 2024
Viewed by 505
Abstract
The growing problem of unsolicited text messages (smishing) and data irregularities necessitates stronger spam detection solutions. This paper explores the development of a sophisticated model designed to identify smishing messages by understanding the complex relationships among words, images, and context-specific factors, areas that [...] Read more.
The growing problem of unsolicited text messages (smishing) and data irregularities necessitates stronger spam detection solutions. This paper explores the development of a sophisticated model designed to identify smishing messages by understanding the complex relationships among words, images, and context-specific factors, areas that remain underexplored in existing research. To address this, we merge a UCI spam dataset of regular text messages with real-world spam data, leveraging OCR technology for comprehensive analysis. The study employs a combination of traditional machine learning models, including K-means, Non-Negative Matrix Factorization, and Gaussian Mixture Models, along with feature extraction techniques such as TF-IDF and PCA. Additionally, deep learning models like RNN-Flatten, LSTM, and Bi-LSTM are utilized. The selection of these models is driven by their complementary strengths in capturing both the linear and non-linear relationships inherent in smishing messages. Machine learning models are chosen for their efficiency in handling structured text data, while deep learning models are selected for their superior ability to capture sequential dependencies and contextual nuances. The performance of these models is rigorously evaluated using metrics like accuracy, precision, recall, and F1 score, enabling a comparative analysis between the machine learning and deep learning approaches. Notably, the K-means feature extraction with vectorizer achieved 91.01% accuracy, and the KNN-Flatten model reached 94.13% accuracy, emerging as the top performer. The rationale behind highlighting these models is their potential to significantly improve smishing detection rates. For instance, the high accuracy of the KNN-Flatten model suggests its applicability in real-time spam detection systems, but its computational complexity might limit scalability in large-scale deployments. Similarly, while K-means with vectorizer excels in accuracy, it may struggle with the dynamic and evolving nature of smishing attacks, necessitating continual retraining. Full article
Show Figures

Figure 1

Back to TopTop