Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (18,150)

Search Parameters:
Keywords = deep learning network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 9272 KiB  
Article
CNN vs. LSTM: A Comparative Study of Hourly Precipitation Intensity Prediction as a Key Factor in Flood Forecasting Frameworks
by Isa Ebtehaj and Hossein Bonakdari
Atmosphere 2024, 15(9), 1082; https://doi.org/10.3390/atmos15091082 - 6 Sep 2024
Abstract
Accurate precipitation intensity forecasting is crucial for effective flood management and early warning systems. This study evaluates the performances of convolutional neural network (CNN) and long short-term memory (LSTM) models in predicting hourly precipitation intensity using data from Sainte Catherine de la Jacques [...] Read more.
Accurate precipitation intensity forecasting is crucial for effective flood management and early warning systems. This study evaluates the performances of convolutional neural network (CNN) and long short-term memory (LSTM) models in predicting hourly precipitation intensity using data from Sainte Catherine de la Jacques Cartier station near Québec City. The models predict precipitation levels from one to six hours ahead, which are categorized into slight, moderate, heavy, and very heavy precipitation intensities. Our methodology involved gathering hourly precipitation data, defining input combinations for multistep ahead forecasting, and employing CNN and LSTM models. The performances of these models were assessed through qualitative and quantitative evaluations. The key findings reveal that the LSTM model excelled in the short-term (1HA to 2HA) and long-term (3HA to 6HA) forecasting, with higher R2 (up to 0.999) and NSE values (up to 0.999), while the CNN model was more computationally efficient, with lower AICc values (e.g., −16,041.1 for 1HA). The error analysis shows that the CNN demonstrated higher precision in the heavy and very heavy categories, with a lower relative error, whereas the LSTM performed better for the slight and moderate categories. The LSTM outperformed the CNN in minor- and high-intensity events, but the CNN exhibited a better performance for significant precipitation events with shorter lead times. Overall, both models were adequate, with the LSTM providing better accuracy for extended forecasts and the CNN offering efficiency for immediate predictions, highlighting their complementary roles in enhancing early warning systems and flood management strategies. Full article
Show Figures

Figure 1

23 pages, 13140 KiB  
Article
MSCR-FuResNet: A Three-Residual Network Fusion Model Based on Multi-Scale Feature Extraction and Enhanced Channel Spatial Features for Close-Range Apple Leaf Diseases Classification under Optimal Conditions
by Xili Chen, Xuanzhu Xing, Yongzhong Zhang, Ruifeng Liu, Lin Li, Ruopeng Zhang, Lei Tang, Ziyang Shi, Hao Zhou, Ruitian Guo and Jingrong Dong
Horticulturae 2024, 10(9), 953; https://doi.org/10.3390/horticulturae10090953 - 6 Sep 2024
Abstract
The precise and automated diagnosis of apple leaf diseases is essential for maximizing apple yield and advancing agricultural development. Despite the widespread utilization of deep learning techniques, several challenges persist: (1) the presence of small disease spots on apple leaves poses difficulties for [...] Read more.
The precise and automated diagnosis of apple leaf diseases is essential for maximizing apple yield and advancing agricultural development. Despite the widespread utilization of deep learning techniques, several challenges persist: (1) the presence of small disease spots on apple leaves poses difficulties for models to capture intricate features; (2) the high similarity among different types of apple leaf diseases complicates their differentiation; and (3) images with complex backgrounds often exhibit low contrast, thereby reducing classification accuracy. To tackle these challenges, we propose a three-residual fusion network known as MSCR-FuResNet (Fusion of Multi-scale Feature Extraction and Enhancements of Channels and Residual Blocks Net), which consists of three sub-networks: (1) enhancing detailed feature extraction through multi-scale feature extraction; (2) improving the discrimination of similar features by suppressing insignificant channels and pixels; and (3) increasing low-contrast feature extraction by modifying the activation function and residual blocks. The model was validated with a comprehensive dataset from public repositories, including Plant Village and Baidu Flying Paddle. Various data augmentation techniques were employed to address class imbalance. Experimental results demonstrate that the proposed model outperforms ResNet-50 with an accuracy of 97.27% on the constructed dataset, indicating significant advancements in apple leaf disease recognition. Full article
(This article belongs to the Section Plant Pathology and Disease Management (PPDM))
Show Figures

Figure 1

21 pages, 558 KiB  
Review
Trackerless 3D Freehand Ultrasound Reconstruction: A Review
by Chrissy A. Adriaans, Mark Wijkhuizen, Lennard M. van Karnenbeek, Freija Geldof and Behdad Dashtbozorg
Appl. Sci. 2024, 14(17), 7991; https://doi.org/10.3390/app14177991 - 6 Sep 2024
Abstract
Two-dimensional ultrasound (2D US) is commonly used in clinical settings for its cost-effectiveness and non-invasiveness, but it is limited by spatial orientation and operator dependency. Three-dimensional ultrasound (3D US) overcomes these limitations by adding a third dimension and enhancing integration with other imaging [...] Read more.
Two-dimensional ultrasound (2D US) is commonly used in clinical settings for its cost-effectiveness and non-invasiveness, but it is limited by spatial orientation and operator dependency. Three-dimensional ultrasound (3D US) overcomes these limitations by adding a third dimension and enhancing integration with other imaging modalities. Advances in deep learning (DL) have further propelled the viability of freehand image-based 3D reconstruction, broadening clinical applications in intraoperative and point-of-care (POC) settings. This review evaluates state-of-the-art freehand 3D US reconstruction methods that eliminate the need for external tracking devices, focusing on experimental setups, data acquisition strategies, and reconstruction methodologies. PubMed, Scopus, and IEEE Xplore were searched for studies since 2014 following the PRISMA guidelines, excluding those using additional imaging or tracking systems other than inertial measurement units (IMUs). Fourteen eligible studies were analyzed, showing a shift from traditional speckle decorrelation towards DL-based methods, particularly convolutional neural networks (CNNs). Variability in datasets and evaluation methods hindered a comprehensive quantitative comparison, but notable accuracy improvements were observed with IMUs and integration of contextual and temporal information within CNNs. These advancements enhance freehand 3D US reconstruction feasibility, though variability limits definitive conclusions about the most effective methods. Future research should focus on improving precision in complex trajectories and adaptability across clinical scenarios. Full article
(This article belongs to the Special Issue Novel Applications of Artificial Intelligence in Ultrasound Imaging)
12 pages, 2853 KiB  
Article
Research on Mitigating Atmosphere Turbulence Fading by Relay Selections in Free-Space Optical Communication Systems with Multi-Transceivers
by Xiaogang San, Zuoyu Liu and Ying Wang
Photonics 2024, 11(9), 847; https://doi.org/10.3390/photonics11090847 - 6 Sep 2024
Abstract
In free-space optical communication (FSOC) systems, atmospheric turbulence can bring about power fluctuations in receiver ends, restricting channel capacity. Relay techniques can divide a long FSOC link into several short links to mitigate the fading events caused by atmospheric turbulence. This paper proposes [...] Read more.
In free-space optical communication (FSOC) systems, atmospheric turbulence can bring about power fluctuations in receiver ends, restricting channel capacity. Relay techniques can divide a long FSOC link into several short links to mitigate the fading events caused by atmospheric turbulence. This paper proposes a Reinforcement Learning-based Relay Selection (RLRS) method based on Deep Q-Network (DQN) in a FSOC system with multiple transceivers, whose aim is to enhance the average channel capacity of the system. Malaga turbulence is studied in this paper. The presence of handover loss is also considered. The relay nodes serve in decode-and-forward (DF). Simulation results demonstrate that the RLRS algorithm outperforms the conventional greedy algorithm, which implies that the RLRS algorithm may be utilized in practical FSOC systems. Full article
(This article belongs to the Special Issue Recent Advances in Optical Turbulence)
Show Figures

Figure 1

33 pages, 23511 KiB  
Article
PARKTag: An AI–Blockchain Integrated Solution for an Efficient, Trusted, and Scalable Parking Management System
by Atharva Kalbhor, Rashmi S. Nair, Shraddha Phansalkar, Rahul Sonkamble, Abhishek Sharma, Harshit Mohan, Chin Hong Wong and Wei Hong Lim
Technologies 2024, 12(9), 155; https://doi.org/10.3390/technologies12090155 - 6 Sep 2024
Abstract
The imbalance between parking availability and demand has led to a rise in traffic challenges in many cities. The adoption of technologies like the Internet of Things and deep learning algorithms has been extensively explored to build automated smart parking systems in urban [...] Read more.
The imbalance between parking availability and demand has led to a rise in traffic challenges in many cities. The adoption of technologies like the Internet of Things and deep learning algorithms has been extensively explored to build automated smart parking systems in urban environments. Non-human-mediated, scalable smart parking systems that are built on decentralized blockchain systems will further enhance transparency and trust in this domain. The presented work, PARKTag, is an integration of a blockchain-based system and computer vision models to detect on-field free parking slots, efficiently navigate vehicles to those slots, and automate the computation of parking fees. This innovative approach aims to enhance the efficiency, scalability, and convenience of parking management by leveraging and integrating advanced technologies for real-time slot detection, navigation, and secure, transparent fee calculation with blockchain smart contracts. PARKTag was evaluated through implementation and emulation in selected areas of the MIT Art Design Technology University campus, with a customized built-in dataset of over 2000 images collected on-field in different conditions. The fine-tuned parking slot detection model leverages pre-trained algorithms and achieves significant performance metrics with a validation accuracy of 92.9% in free slot detection. With the Solidity smart contract deployed on the Ethereum test network, PARKTag achieved a significant throughput of 10 user requests per second in peak traffic hours. PARKTag is implemented as a mobile application and deployed in the mobile application store. Its beta version has undergone user validation for feedback and acceptance, marking a significant step toward the development of the final product. Full article
Show Figures

Figure 1

18 pages, 3237 KiB  
Article
Lightweight Wheat Spike Detection Method Based on Activation and Loss Function Enhancements for YOLOv5s
by Jingsong Li, Feijie Dai, Haiming Qian, Linsheng Huang and Jinling Zhao
Agronomy 2024, 14(9), 2036; https://doi.org/10.3390/agronomy14092036 - 6 Sep 2024
Abstract
Wheat spike count is one of the critical indicators for assessing the growth and yield of wheat. However, illumination variations, mutual occlusion, and background interference have greatly affected wheat spike detection. A lightweight detection method was proposed based on the YOLOv5s. Initially, the [...] Read more.
Wheat spike count is one of the critical indicators for assessing the growth and yield of wheat. However, illumination variations, mutual occlusion, and background interference have greatly affected wheat spike detection. A lightweight detection method was proposed based on the YOLOv5s. Initially, the original YOLOv5s was improved by combing the additional small-scale detection layer and integrating the ECA (Efficient Channel Attention) attention mechanism into all C3 modules (YOLOv5s + 4 + ECAC3). After comparing GhostNet, ShuffleNetV2, and MobileNetV3, the GhostNet architecture was finally selected as the optimal lightweight model framework based on its superior performance in various evaluations. Subsequently, the incorporation of five different activation functions into the network led to the identification of the RReLU (Randomized Leaky ReLU) activation function as the most effective in augmenting the network’s performance. Ultimately, the network’s loss function of CIoU (Complete Intersection over Union) was optimized using the EIoU (Efficient Intersection over Union) loss function. Despite a minor reduction of 2.17% in accuracy for the refined YOLOv5s + 4 + ECAC3 + G + RR + E network when compared to the YOLOv5s + 4 + ECAC3, there was a marginal improvement of 0.77% over the original YOLOv5s. Furthermore, the parameter count was diminished by 32% and 28.2% relative to the YOLOv5s + 4 + ECAC3 and YOLOv5s, respectively. The model size was reduced by 28.0% and 20%, and the Giga Floating-point Operations Per Second (GFLOPs) were lowered by 33.2% and 9.5%, respectively, signifying a substantial improvement in the network’s efficiency without significantly compromising accuracy. This study offers a methodological reference for the rapid and accurate detection of agricultural objects through the enhancement of a deep learning network. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

18 pages, 4881 KiB  
Article
An Improved Deep Deterministic Policy Gradient Pantograph Active Control Strategy for High-Speed Railways
by Ying Wang, Yuting Wang, Xiaoqiang Chen, Yixuan Wang and Zhanning Chang
Electronics 2024, 13(17), 3545; https://doi.org/10.3390/electronics13173545 - 6 Sep 2024
Abstract
The pantograph–catenary system (PCS) is essential for trains to obtain electrical energy. As the train’s operating speed increases, the vibration between the pantograph and the catenary intensifies, reducing the quality of the current collection. Active control may significantly reduce the vibration of the [...] Read more.
The pantograph–catenary system (PCS) is essential for trains to obtain electrical energy. As the train’s operating speed increases, the vibration between the pantograph and the catenary intensifies, reducing the quality of the current collection. Active control may significantly reduce the vibration of the PCS, effectively lower the cost of line retrofitting, and enhance the quality of the current collection. This article proposes an improved deep deterministic policy gradient (IDDPG) for the pantograph active control problem, which delays updating the Actor and Target–Actor networks and adopts a reconstructed experience replay mechanism. The deep reinforcement learning (DRL) environment module was first established by creating a PCS coupling model. On this basis, the controller’s DRL module is precisely designed using the IDDPG strategy. Ultimately, the control strategy is integrated with the PCS for training, and the controller’s performance is validated on the PCS. Simulation experiments show that the improved strategy significantly reduces the training time, enhances the steady-state performance of the agent during later training stages, and effectively reduces the standard deviation of the pantograph–catenary contact force (PCCF) by an average of over 51.44%, effectively improving the quality of current collection. Full article
Show Figures

Figure 1

39 pages, 6629 KiB  
Article
A Combined CNN Architecture for Speech Emotion Recognition
by Rolinson Begazo, Ana Aguilera, Irvin Dongo and Yudith Cardinale
Sensors 2024, 24(17), 5797; https://doi.org/10.3390/s24175797 - 6 Sep 2024
Abstract
Emotion recognition through speech is a technique employed in various scenarios of Human–Computer Interaction (HCI). Existing approaches have achieved significant results; however, limitations persist, with the quantity and diversity of data being more notable when deep learning techniques are used. The lack of [...] Read more.
Emotion recognition through speech is a technique employed in various scenarios of Human–Computer Interaction (HCI). Existing approaches have achieved significant results; however, limitations persist, with the quantity and diversity of data being more notable when deep learning techniques are used. The lack of a standard in feature selection leads to continuous development and experimentation. Choosing and designing the appropriate network architecture constitutes another challenge. This study addresses the challenge of recognizing emotions in the human voice using deep learning techniques, proposing a comprehensive approach, and developing preprocessing and feature selection stages while constructing a dataset called EmoDSc as a result of combining several available databases. The synergy between spectral features and spectrogram images is investigated. Independently, the weighted accuracy obtained using only spectral features was 89%, while using only spectrogram images, the weighted accuracy reached 90%. These results, although surpassing previous research, highlight the strengths and limitations when operating in isolation. Based on this exploration, a neural network architecture composed of a CNN1D, a CNN2D, and an MLP that fuses spectral features and spectogram images is proposed. The model, supported by the unified dataset EmoDSc, demonstrates a remarkable accuracy of 96%. Full article
(This article belongs to the Special Issue Emotion Recognition Based on Sensors (3rd Edition))
Show Figures

Figure 1

24 pages, 8893 KiB  
Article
Assessing Data Preparation and Machine Learning for Tree Species Classification Using Hyperspectral Imagery
by Wenge Ni-Meister, Anthony Albanese and Francesca Lingo
Remote Sens. 2024, 16(17), 3313; https://doi.org/10.3390/rs16173313 - 6 Sep 2024
Abstract
Tree species classification using hyperspectral imagery shows incredible promise in developing a large-scale, high-resolution model for identifying tree species, providing unprecedented details on global tree species distribution. Many questions remain unanswered about the best practices for creating a global, general hyperspectral tree species [...] Read more.
Tree species classification using hyperspectral imagery shows incredible promise in developing a large-scale, high-resolution model for identifying tree species, providing unprecedented details on global tree species distribution. Many questions remain unanswered about the best practices for creating a global, general hyperspectral tree species classification model. This study aims to address three key issues in creating a hyperspectral species classification model. We assessed the effectiveness of three data-labeling methods to create training data, three data-splitting methods for training/validation/testing, and machine-learning and deep-learning (including semi-supervised deep-learning) models for tree species classification using hyperspectral imagery at National Ecological Observatory Network (NEON) Sites. Our analysis revealed that the existing data-labeling method using the field vegetation structure survey performed reasonably well. The random tree data-splitting technique was the most efficient method for both intra-site and inter-site classifications to overcome the impact of spatial autocorrelation to avoid the potential to create a locally overfit model. Deep learning consistently outperformed random forest classification; both semi-supervised and supervised deep-learning models displayed the most promising results in creating a general taxa-classification model. This work has demonstrated the possibility of developing tree-classification models that can identify tree species from outside their training area and that semi-supervised deep learning may potentially utilize the untapped terabytes of unlabeled forest imagery. Full article
Show Figures

Figure 1

43 pages, 3605 KiB  
Review
In-Depth Insights into the Application of Recurrent Neural Networks (RNNs) in Traffic Prediction: A Comprehensive Review
by Yuxin He, Ping Huang, Weihang Hong, Qin Luo, Lishuai Li and Kwok-Leung Tsui
Algorithms 2024, 17(9), 398; https://doi.org/10.3390/a17090398 - 6 Sep 2024
Abstract
Traffic prediction is crucial for transportation management and user convenience. With the rapid development of deep learning techniques, numerous models have emerged for traffic prediction. Recurrent Neural Networks (RNNs) are extensively utilized as representative predictive models in this domain. This paper comprehensively reviews [...] Read more.
Traffic prediction is crucial for transportation management and user convenience. With the rapid development of deep learning techniques, numerous models have emerged for traffic prediction. Recurrent Neural Networks (RNNs) are extensively utilized as representative predictive models in this domain. This paper comprehensively reviews RNN applications in traffic prediction, focusing on their significance and challenges. The review begins by discussing the evolution of traffic prediction methods and summarizing state-of-the-art techniques. It then delves into the unique characteristics of traffic data, outlines common forms of input representations in traffic prediction, and generalizes an abstract description of traffic prediction problems. Then, the paper systematically categorizes models based on RNN structures designed for traffic prediction. Moreover, it provides a comprehensive overview of seven sub-categories of applications of deep learning models based on RNN in traffic prediction. Finally, the review compares RNNs with other state-of-the-art methods and highlights the challenges RNNs face in traffic prediction. This review is expected to offer significant reference value for comprehensively understanding the various applications of RNNs and common state-of-the-art models in traffic prediction. By discussing the strengths and weaknesses of these models and proposing strategies to address the challenges faced by RNNs, it aims to provide scholars with insights for designing better traffic prediction models. Full article
(This article belongs to the Section Algorithms for Multidisciplinary Applications)
Show Figures

Figure 1

26 pages, 8282 KiB  
Article
Advanced Fault Detection in Power Transformers Using Improved Wavelet Analysis and LSTM Networks Considering Current Transformer Saturation and Uncertainties
by Qusay Alhamd, Mohsen Saniei, Seyyed Ghodratollah Seifossadat and Elaheh Mashhour
Algorithms 2024, 17(9), 397; https://doi.org/10.3390/a17090397 - 6 Sep 2024
Abstract
Power transformers are vital and costly components in power systems, essential for ensuring a reliable and uninterrupted supply of electrical energy. Their protection is crucial for improving reliability, maintaining network stability, and minimizing operational costs. Previous studies have introduced differential protection schemes with [...] Read more.
Power transformers are vital and costly components in power systems, essential for ensuring a reliable and uninterrupted supply of electrical energy. Their protection is crucial for improving reliability, maintaining network stability, and minimizing operational costs. Previous studies have introduced differential protection schemes with harmonic restraint to detect internal transformer faults. However, these schemes often struggle with computational inaccuracies in fault detection due to neglecting current transformer (CT) saturation and associated uncertainties. CT saturation during internal faults can produce even harmonics, disrupting relay operations. Additionally, CT saturation during transformer energization can introduce a DC component, leading to incorrect relay activation. This paper introduces a novel feature extracted through advanced wavelet transform analysis of differential current. This feature, combined with differential current amplitude and bias current, is used to train a deep learning system based on long short-term memory (LSTM) networks. By accounting for existing uncertainties, this system accurately identifies internal transformer faults under various CT saturation and measurement uncertainty conditions. Test and validation results demonstrate the proposed method’s effectiveness and superiority in detecting internal faults in power transformers, even in the presence of CT saturation, outperforming other recent modern techniques. Full article
Show Figures

Figure 1

18 pages, 3644 KiB  
Article
Edge Detection in Colored Images Using Parallel CNNs and Social Spider Optimization
by Jiahao Zhang, Wei Wang and Jianfei Wang
Electronics 2024, 13(17), 3540; https://doi.org/10.3390/electronics13173540 - 6 Sep 2024
Abstract
Edge detection is a crucial issue in computer vision, with convolutional neural networks (CNNs) being a key component in various systems for detecting edges within images, offering numerous practical implementations. This paper introduces a hybrid approach for edge detection in color images using [...] Read more.
Edge detection is a crucial issue in computer vision, with convolutional neural networks (CNNs) being a key component in various systems for detecting edges within images, offering numerous practical implementations. This paper introduces a hybrid approach for edge detection in color images using an enhanced holistically led edge detection (HED) structure. The method consists of two primary phases: edge approximation based on parallel convolutional neural networks (PCNNs) and edge enhancement based on social spider optimization (SSO). The first phase uses two parallel CNN models to preliminarily approximate image edges. The first model uses edge-detected images from the Otsu-Canny operator, while the second model accepts RGB color images as input. The output of the proposed PCNN model is compared with pairwise combination of color layers in the input image. In the second phase, the SSO algorithm is used to optimize the edge detection result, modifying edges in the approximate image to minimize differences with the resulting color layer combinations. The experimental results demonstrate that our proposed method achieved a precision of 0.95. Furthermore, the peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) values stand at 20.39 and 0.83, respectively. The high PSNR value of our method signifies superior output quality, showing reduced contrast and noise compared to the ground truth image. Similarly, the SSIM value indicates that the method’s edge structure surpasses that of the ground truth image, further affirming its superiority over other methods. Full article
Show Figures

Figure 1

19 pages, 6430 KiB  
Article
An Ensemble Deep Neural Network-Based Method for Person Identification Using Electrocardiogram Signals Acquired on Different Days
by Yeong-Hyeon Byeon and Keun-Chang Kwak
Appl. Sci. 2024, 14(17), 7959; https://doi.org/10.3390/app14177959 - 6 Sep 2024
Viewed by 53
Abstract
Electrocardiogram (ECG) signals are a measure minute electrical signals generated during the cardiac cycle, a biometric signal that occurs during vital human activity. ECG signals are susceptible to various types of noise depending on the data acquisition conditions, with factors such as sensor [...] Read more.
Electrocardiogram (ECG) signals are a measure minute electrical signals generated during the cardiac cycle, a biometric signal that occurs during vital human activity. ECG signals are susceptible to various types of noise depending on the data acquisition conditions, with factors such as sensor placement and the physiological and mental states of the subject contributing to the diverse shapes of these signals. When the data are acquired in a single session, the environmental variables are relatively similar, resulting in similar ECG signals; however, in subsequent sessions, even for the same person, changes in the environmental variables can alter the signal shape. This phenomenon poses challenges for person identification using ECG signals acquired on different days. To improve the performance of individual identification, even when ECG data is acquired on different days, this paper proposes an ensemble deep neural network for person identification by comparing and analyzing the ECG recognition performance under various conditions. The proposed ensemble deep neural network comprises three streams that incorporate two well-known pretrained models. Each network receives the time-frequency representation of ECG signals as input, and a stream reuses the same network structure under different learning conditions with or without data augmentation. The proposed ensemble deep neural network was validated on the Physikalisch-Technische Bundesanstalt dataset, and the results confirmed a 3.39% improvement in accuracy compared to existing methods. Full article
(This article belongs to the Section Applied Biosciences and Bioengineering)
Show Figures

Figure 1

15 pages, 4276 KiB  
Article
Spectrum Sensing Method Based on STFT-RADN in Cognitive Radio Networks
by Anyi Wang, Tao Zhu and Qifeng Meng
Sensors 2024, 24(17), 5792; https://doi.org/10.3390/s24175792 - 6 Sep 2024
Viewed by 69
Abstract
To address the common issues in traditional convolutional neural network (CNN)-based spectrum sensing algorithms in cognitive radio networks (CRNs), including inadequate signal feature representation, inefficient utilization of feature map information, and limited feature extraction capabilities due to shallow network structures, this paper proposes [...] Read more.
To address the common issues in traditional convolutional neural network (CNN)-based spectrum sensing algorithms in cognitive radio networks (CRNs), including inadequate signal feature representation, inefficient utilization of feature map information, and limited feature extraction capabilities due to shallow network structures, this paper proposes a spectrum sensing algorithm based on a short-time Fourier transform (STFT) and residual attention dense network (RADN). Specifically, the RADN model improves the basic residual block and introduces the convolutional block attention module (CBAM), combining residual connections and dense connections to form a powerful deep feature extraction structure known as residual in dense (RID). This significantly enhances the network’s feature extraction capabilities. By performing STFT on the received signals and normalizing them, the signals are converted into time–frequency spectrograms as network inputs, better capturing signal features. The RADN is trained to extract abstract features from the time–frequency images, and the trained RADN serves as the final classifier for spectrum sensing. Experimental results demonstrate that the STFT-RADN spectrum sensing method significantly improves performance under low signal-to-noise ratio (SNR) conditions compared to traditional deep-learning-based methods. This method not only adapts to various modulation schemes but also exhibits high detection probability and strong robustness. Full article
(This article belongs to the Special Issue Sensors for Enabling Wireless Spectrum Access)
Show Figures

Figure 1

20 pages, 677 KiB  
Article
MAMGD: Gradient-Based Optimization Method Using Exponential Decay
by Nikita Sakovich, Dmitry Aksenov, Ekaterina Pleshakova and Sergey Gataullin
Technologies 2024, 12(9), 154; https://doi.org/10.3390/technologies12090154 - 6 Sep 2024
Viewed by 96
Abstract
Optimization methods, namely, gradient optimization methods, are a key part of neural network training. In this paper, we propose a new gradient optimization method using exponential decay and the adaptive learning rate using a discrete second-order derivative of gradients. The MAMGD optimizer uses [...] Read more.
Optimization methods, namely, gradient optimization methods, are a key part of neural network training. In this paper, we propose a new gradient optimization method using exponential decay and the adaptive learning rate using a discrete second-order derivative of gradients. The MAMGD optimizer uses an adaptive learning step, exponential smoothing and gradient accumulation, parameter correction, and some discrete analogies from classical mechanics. The experiments included minimization of multivariate real functions, function approximation using multilayer neural networks, and training neural networks on popular classification and regression datasets. The experimental results of the new optimization technology showed a high convergence speed, stability to fluctuations, and an accumulation of gradient accumulators. The research methodology is based on the quantitative performance analysis of the algorithm by conducting computational experiments on various optimization problems and comparing it with existing methods. Full article
Show Figures

Figure 1

Back to TopTop