Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (11,646)

Search Parameters:
Keywords = CNN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
9 pages, 4309 KiB  
Communication
Attention Mechanism-Based Glaucoma Classification Model Using Retinal Fundus Images
by You-Sang Cho, Ho-Jung Song, Ju-Hyuck Han and Yong-Suk Kim
Sensors 2024, 24(14), 4684; https://doi.org/10.3390/s24144684 (registering DOI) - 19 Jul 2024
Abstract
This paper presents a classification model for eye diseases utilizing attention mechanisms to learn features from fundus images and structures. The study focuses on diagnosing glaucoma by extracting retinal vessels and the optic disc from fundus images using a ResU-Net-based segmentation model and [...] Read more.
This paper presents a classification model for eye diseases utilizing attention mechanisms to learn features from fundus images and structures. The study focuses on diagnosing glaucoma by extracting retinal vessels and the optic disc from fundus images using a ResU-Net-based segmentation model and Hough Circle Transform, respectively. The extracted structures and preprocessed images were inputted into a CNN-based multi-input model for training. Comparative evaluations demonstrated that our model outperformed other research models in classifying glaucoma, even with a smaller dataset. Ablation studies confirmed that using attention mechanisms to learn fundus structures significantly enhanced performance. The study also highlighted the challenges in normal case classification due to potential feature degradation during structure extraction. Future research will focus on incorporating additional fundus structures such as the macula, refining extraction algorithms, and expanding the types of classified eye diseases. Full article
Show Figures

Figure 1

14 pages, 4781 KiB  
Article
A 1D Convolutional Neural Network (1D-CNN) Temporal Filter for Atmospheric Variability: Reducing the Sensitivity of Filtering Accuracy to Missing Data Points
by Dan Yu, Hoiio Kong, Jeremy Cheuk-Hin Leung, Pak Wai Chan, Clarence Fong, Yuchen Wang and Banglin Zhang
Appl. Sci. 2024, 14(14), 6289; https://doi.org/10.3390/app14146289 (registering DOI) - 19 Jul 2024
Abstract
The atmosphere exhibits variability across different time scales. Currently, in the field of atmospheric science, statistical filtering is one of the most widely used methods for extracting signals on certain time scales. However, signal extraction based on traditional statistical filters may be sensitive [...] Read more.
The atmosphere exhibits variability across different time scales. Currently, in the field of atmospheric science, statistical filtering is one of the most widely used methods for extracting signals on certain time scales. However, signal extraction based on traditional statistical filters may be sensitive to missing data points, which are particularly common in meteorological data. To address this issue, this study applies a new type of temporal filters based on a one-dimensional convolution neural network (1D-CNN) and examines its performance on reducing such uncertainties. As an example, we investigate the advantages of a 1D-CNN bandpass filter in extracting quasi-biweekly-to-intraseasonal signals (10–60 days) from temperature data provided by the Hong Kong Observatory. The results show that the 1D-CNN achieves accuracies similar to a 121-point Lanczos filter. In addition, the 1D-CNN filter allows a maximum of 10 missing data points within the 60-point window length, while keeping its accuracy higher than 80% (R2 > 0.8). This indicates that the 1D-CNN model works well even when missing data points exist in the time series. This study highlights another potential for applying machine learning algorithms in atmospheric and climate research, which will be useful for future research involving incomplete time series and real-time filtering. Full article
(This article belongs to the Special Issue Machine Learning Approaches for Geophysical Data Analysis)
Show Figures

Figure 1

12 pages, 1800 KiB  
Article
Research on Public Service Request Text Classification Based on BERT-BiLSTM-CNN Feature Fusion
by Yunpeng Xiong, Guolian Chen and Junkuo Cao
Appl. Sci. 2024, 14(14), 6282; https://doi.org/10.3390/app14146282 (registering DOI) - 18 Jul 2024
Viewed by 82
Abstract
Convolutional neural networks (CNNs) face challenges in capturing long-distance text correlations, and Bidirectional Long Short-Term Memory (BiLSTM) networks exhibit limited feature extraction capabilities for text classification of public service requests. To address the abovementioned problems, this work utilizes an ensemble learning approach to [...] Read more.
Convolutional neural networks (CNNs) face challenges in capturing long-distance text correlations, and Bidirectional Long Short-Term Memory (BiLSTM) networks exhibit limited feature extraction capabilities for text classification of public service requests. To address the abovementioned problems, this work utilizes an ensemble learning approach to integrate model elements efficiently. This study presents a method for classifying public service request text using a hybrid neural network model called BERT-BiLSTM-CNN. First, BERT (Bidirectional Encoder Representations from Transformers) is used for preprocessing to obtain text vector representations. Then, context and process sequence information are captured through BiLSTM. Next, local features in the text are captured through CNN. Finally, classification results are obtained through Softmax. Through comparative analysis, the method of fusing these three models is superior to other hybrid neural network model architectures in multiple classification tasks. It has a significant effect on public service request text classification. Full article
Show Figures

Figure 1

20 pages, 5032 KiB  
Article
Enhanced Learning Enriched Features Mechanism Using Deep Convolutional Neural Network for Image Denoising and Super-Resolution
by Iqra Waseem, Muhammad Habib, Eid Rehman, Ruqia Bibi, Rehan Mehmood Yousaf, Muhammad Aslam, Syeda Fizzah Jilani and Muhammad Waqar Younis
Appl. Sci. 2024, 14(14), 6281; https://doi.org/10.3390/app14146281 - 18 Jul 2024
Viewed by 70
Abstract
Image denoising and super-resolution play vital roles in imaging systems, greatly reducing the preprocessing cost of many AI techniques for object detection, segmentation, and tracking. Various advancements have been accomplished in this field, but progress is still needed. In this paper, we have [...] Read more.
Image denoising and super-resolution play vital roles in imaging systems, greatly reducing the preprocessing cost of many AI techniques for object detection, segmentation, and tracking. Various advancements have been accomplished in this field, but progress is still needed. In this paper, we have proposed a novel technique named the Enhanced Learning Enriched Features (ELEF) mechanism using a deep convolutional neural network, which makes significant improvements to existing techniques. ELEF consists of two major processes: (1) Denoising, which removes the noise from images; and (2) Super-resolution, which improves the clarity and details of images. Features are learned through deep CNN and not through traditional algorithms so that we can better refine and enhance images. To effectively capture features, the network architecture adopted Dual Attention Units (DUs), which align with the Multi-Scale Residual Block (MSRB) for robust feature extraction, working sidewise with the feature-matching Selective Kernel Extraction (SKF). In addition, resolution mismatching cases are processed in detail to produce high-quality images. The effectiveness of the ELEF model is highlighted by the performance metrics, achieving a Peak Signal-to-Noise Ratio (PSNR) of 42.99 and a Structural Similarity Index (SSIM) of 0.9889, which indicates the ability to carry out the desired high-quality image restoration and enhancement. Full article
(This article belongs to the Special Issue Advances in Image Enhancement and Restoration Technology)
Show Figures

Figure 1

18 pages, 1603 KiB  
Article
SPT-UNet: A Superpixel-Level Feature Fusion Network for Water Extraction from SAR Imagery
by Teng Zhao, Xiaoping Du, Chen Xu, Hongdeng Jian, Zhipeng Pei, Junjie Zhu, Zhenzhen Yan and Xiangtao Fan
Remote Sens. 2024, 16(14), 2636; https://doi.org/10.3390/rs16142636 - 18 Jul 2024
Viewed by 85
Abstract
Extracting water bodies from synthetic aperture radar (SAR) images plays a crucial role in the management of water resources, flood monitoring, and other applications. Recently, transformer-based models have been extensively utilized in the remote sensing domain. However, due to regular patch-partition and weak [...] Read more.
Extracting water bodies from synthetic aperture radar (SAR) images plays a crucial role in the management of water resources, flood monitoring, and other applications. Recently, transformer-based models have been extensively utilized in the remote sensing domain. However, due to regular patch-partition and weak inductive bias, transformer-based models face challenges such as edge serration and high data dependency when used for water body extraction from SAR images. To address these challenges, we introduce a new model, the Superpixel-based Transformer (SPT), based on the adaptive characteristic of superpixels and knowledge constraints of the adjacency matrix. (1) To mitigate edge serration, the SPT replaces regular patch partition with superpixel segmentation to fully utilize the internal homogeneity of superpixels. (2) To reduce data dependency, the SPT incorporates a normalized adjacency matrix between superpixels into the Multi-Layer Perceptron (MLP) to impose knowledge constraints. (3) Additionally, to integrate superpixel-level learning from the SPT with pixel-level learning from the CNN, we combine these two deep networks to form SPT-UNet for water body extraction. The results show that our SPT-UNet is competitive compared with other state-of-the-art extraction models, both in terms of quantitative metrics and visual effects. Full article
40 pages, 29439 KiB  
Article
A Multivariate Time Series Prediction Method Based on Convolution-Residual Gated Recurrent Neural Network and Double-Layer Attention
by Chuxin Cao, Jianhong Huang, Man Wu, Zhizhe Lin and Yan Sun
Electronics 2024, 13(14), 2834; https://doi.org/10.3390/electronics13142834 - 18 Jul 2024
Viewed by 108
Abstract
In multivariate and multistep time series prediction research, we often face the problems of insufficient spatial feature extraction and insufficient time-dependent mining of historical series data, which also brings great challenges to multivariate time series analysis and prediction. Inspired by the attention mechanism [...] Read more.
In multivariate and multistep time series prediction research, we often face the problems of insufficient spatial feature extraction and insufficient time-dependent mining of historical series data, which also brings great challenges to multivariate time series analysis and prediction. Inspired by the attention mechanism and residual module, this study proposes a multivariate time series prediction method based on a convolutional-residual gated recurrent hybrid model (CNN-DA-RGRU) with a two-layer attention mechanism to solve the multivariate time series prediction problem in these two stages. Specifically, the convolution module of the proposed model is used to extract the relational features among the sequences, and the two-layer attention mechanism can pay more attention to the relevant variables and give them higher weights to eliminate the irrelevant features, while the residual gated loop module is used to extract the time-varying features of the sequences, in which the residual block is used to achieve the direct connectivity to enhance the expressive power of the model, to solve the gradient explosion and vanishing scenarios, and to facilitate gradient propagation. Experiments were conducted on two public datasets using the proposed model to determine the model hyperparameters, and ablation experiments were conducted to verify the effectiveness of the model; by comparing it with several models, the proposed model was found to achieve good results in multivariate time series-forecasting tasks. Full article
(This article belongs to the Special Issue Machine Learning in Data Analytics and Prediction)
Show Figures

Figure 1

12 pages, 1944 KiB  
Article
Deep Learning-Based Automated Cell Detection-Facilitated Meat Quality Evaluation
by Hui Zheng, Nan Zhao, Saifei Xu, Jin He, Ricardo Ospina, Zhengjun Qiu and Yufei Liu
Foods 2024, 13(14), 2270; https://doi.org/10.3390/foods13142270 - 18 Jul 2024
Viewed by 107
Abstract
Meat consumption is increasing globally. The safety and quality of meat are considered important issues for human health. During evaluations of meat quality and freshness, microbiological parameters are often analyzed. Counts of indicator cells can provide important references for meat quality. In order [...] Read more.
Meat consumption is increasing globally. The safety and quality of meat are considered important issues for human health. During evaluations of meat quality and freshness, microbiological parameters are often analyzed. Counts of indicator cells can provide important references for meat quality. In order to eliminate the error of manual operation and improve detection efficiency, this paper proposed a Convolutional Neural Network (CNN) with a backbone called Detect-Cells-Rapidly-Net (DCRNet), which can identify and count stained cells automatically. The DCRNet replaces the single channel of residual blocks with the aggregated residual blocks to learn more features with fewer parameters. The DCRNet combines the deformable convolution network to fit flexible shapes of stained animal cells. The proposed CNN with DCRNet is self-adaptive to different resolutions of images. The experimental results indicate that the proposed CNN with DCRNet achieves an Average Precision of 81.2% and is better than traditional neural networks for this task. The difference between the results of the proposed method and manual counting is less than 0.5% of the total number of cells. The results indicate that DCRNet is a promising solution for cell detection and can be equipped in future meat quality monitoring systems. Full article
17 pages, 7982 KiB  
Article
Deep Dynamic Weights for Underwater Image Restoration
by Hafiz Shakeel Ahmad Awan and Muhammad Tariq Mahmood
J. Mar. Sci. Eng. 2024, 12(7), 1208; https://doi.org/10.3390/jmse12071208 - 18 Jul 2024
Viewed by 102
Abstract
Underwater imaging presents unique challenges, notably color distortions and reduced contrast due to light attenuation and scattering. Most underwater image enhancement methods first use linear transformations for color compensation and then enhance the image. We observed that linear transformation for color compensation is [...] Read more.
Underwater imaging presents unique challenges, notably color distortions and reduced contrast due to light attenuation and scattering. Most underwater image enhancement methods first use linear transformations for color compensation and then enhance the image. We observed that linear transformation for color compensation is not suitable for certain images. For such images, non-linear mapping is a better choice. This paper introduces a unique underwater image restoration approach leveraging a streamlined convolutional neural network (CNN) for dynamic weight learning for linear and non-linear mapping. In the first phase, a classifier is applied that classifies the input images as Type I or Type II. In the second phase, we use the Deep Line Model (DLM) for Type-I images and the Deep Curve Model (DCM) for Type-II images. For mapping an input image to an output image, the DLM creatively combines color compensation and contrast adjustment in a single step and uses deep lines for transformation, whereas the DCM employs higher-order curves. Both models utilize lightweight neural networks that learn per-pixel dynamic weights based on the input image’s characteristics. Comprehensive evaluations on benchmark datasets using metrics like peak signal-to-noise ratio (PSNR) and root mean square error (RMSE) affirm our method’s effectiveness in accurately restoring underwater images, outperforming existing techniques. Full article
(This article belongs to the Special Issue Application of Deep Learning in Underwater Image Processing)
Show Figures

Figure 1

20 pages, 5228 KiB  
Article
Remote Sensing Image Change Detection Based on Deep Learning: Multi-Level Feature Cross-Fusion with 3D-Convolutional Neural Networks
by Sibo Yu, Chen Tao, Guang Zhang, Yubo Xuan and Xiaodong Wang
Appl. Sci. 2024, 14(14), 6269; https://doi.org/10.3390/app14146269 (registering DOI) - 18 Jul 2024
Viewed by 125
Abstract
Change detection (CD) in high-resolution remote sensing imagery remains challenging due to the complex nature of objects and varying spectral characteristics across different times and locations. Convolutional neural networks (CNNs) have shown promising performance in CD tasks by extracting meaningful semantic features. However, [...] Read more.
Change detection (CD) in high-resolution remote sensing imagery remains challenging due to the complex nature of objects and varying spectral characteristics across different times and locations. Convolutional neural networks (CNNs) have shown promising performance in CD tasks by extracting meaningful semantic features. However, traditional 2D-CNNs may struggle to accurately integrate deep features from multi-temporal images, limiting their ability to improve CD accuracy. This study proposes a Multi-level Feature Cross-Fusion (MFCF) network with 3D-CNNs for remote sensing image change detection. The network aims to effectively extract and fuse deep features from multi-temporal images to identify surface changes. To bridge the semantic gap between high-level and low-level features, a MFCF module is introduced. A channel attention mechanism (CAM) is also integrated to enhance model performance, interpretability, and generalization capabilities. The proposed methodology is validated on the LEVIR construction dataset (LEVIR-CD). The experimental results demonstrate superior performance compared to the current state-of-the-art in evaluation metrics including recall, F1 score, and IOU. The MFCF network, which combines 3D-CNNs and a CAM, effectively utilizes multi-temporal information and deep feature fusion, resulting in precise and reliable change detection in remote sensing imagery. This study significantly contributes to the advancement of change detection methods, facilitating more efficient management and decision making across various domains such as urban planning, natural resource management, and environmental monitoring. Full article
(This article belongs to the Special Issue Advances in Image Recognition and Processing Technologies)
Show Figures

Figure 1

21 pages, 3747 KiB  
Article
ViT-PSO-SVM: Cervical Cancer Predication Based on Integrating Vision Transformer with Particle Swarm Optimization and Support Vector Machine
by Abdulaziz AlMohimeed, Mohamed Shehata, Nora El-Rashidy, Sherif Mostafa, Amira Samy Talaat and Hager Saleh
Bioengineering 2024, 11(7), 729; https://doi.org/10.3390/bioengineering11070729 - 18 Jul 2024
Viewed by 111
Abstract
Cervical cancer (CCa) is the fourth most prevalent and common cancer affecting women worldwide, with increasing incidence and mortality rates. Hence, early detection of CCa plays a crucial role in improving outcomes. Non-invasive imaging procedures with good diagnostic performance are desirable and have [...] Read more.
Cervical cancer (CCa) is the fourth most prevalent and common cancer affecting women worldwide, with increasing incidence and mortality rates. Hence, early detection of CCa plays a crucial role in improving outcomes. Non-invasive imaging procedures with good diagnostic performance are desirable and have the potential to lessen the degree of intervention associated with the gold standard, biopsy. Recently, artificial intelligence-based diagnostic models such as Vision Transformers (ViT) have shown promising performance in image classification tasks, rivaling or surpassing traditional convolutional neural networks (CNNs). This paper studies the effect of applying a ViT to predict CCa using different image benchmark datasets. A newly developed approach (ViT-PSO-SVM) was presented for boosting the results of the ViT based on integrating the ViT with particle swarm optimization (PSO), and support vector machine (SVM). First, the proposed framework extracts features from the Vision Transformer. Then, PSO is used to reduce the complexity of extracted features and optimize feature representation. Finally, a softmax classification layer is replaced with an SVM classification model to precisely predict CCa. The models are evaluated using two benchmark cervical cell image datasets, namely SipakMed and Herlev, with different classification scenarios: two, three, and five classes. The proposed approach achieved 99.112% accuracy and 99.113% F1-score for SipakMed with two classes and achieved 97.778% accuracy and 97.805% F1-score for Herlev with two classes outperforming other Vision Transformers, CNN models, and pre-trained models. Finally, GradCAM is used as an explainable artificial intelligence (XAI) tool to visualize and understand the regions of a given image that are important for a model’s prediction. The obtained experimental results demonstrate the feasibility and efficacy of the developed ViT-PSO-SVM approach and hold the promise of providing a robust, reliable, accurate, and non-invasive diagnostic tool that will lead to improved healthcare outcomes worldwide. Full article
Show Figures

Graphical abstract

20 pages, 560 KiB  
Article
Deep Learning Soft-Decision GNSS Multipath Detection and Mitigation
by Fernando Nunes and Fernando Sousa
Sensors 2024, 24(14), 4663; https://doi.org/10.3390/s24144663 - 18 Jul 2024
Viewed by 117
Abstract
A technique is proposed to detect the presence of the multipath effect in Global Navigation Satellite Signal (GNSS) signals using a convolutional neural network (CNN) as the building block. The network is trained and validated, for a wide range of [...] Read more.
A technique is proposed to detect the presence of the multipath effect in Global Navigation Satellite Signal (GNSS) signals using a convolutional neural network (CNN) as the building block. The network is trained and validated, for a wide range of C/N0 values, with a realistic dataset constituted by the synthetic noisy outputs of a 2D grid of correlators associated with different Doppler frequencies and code delays (time-domain dataset). Multipath-disturbed signals are generated in agreement with the various scenarios encompassed by the adopted multipath model. It was found that pre-processing the outputs of the correlators grid with the two-dimensional Discrete Fourier Transform (frequency-domain dataset) enables the CNN to improve the accuracy relative to the time-domain dataset. Depending on the kind of CNN outputs, two strategies can then be devised to solve the equation of navigation: either remove the disturbed signal from the equation (hard decision) or process the pseudoranges with a weighted least-squares algorithm, where the entries of the weighting matrix are computed using the analog outputs of the neural network (soft decision). Full article
Show Figures

Figure 1

17 pages, 5024 KiB  
Article
SCAE—Stacked Convolutional Autoencoder for Fault Diagnosis of a Hydraulic Piston Pump with Limited Data Samples
by Oybek Eraliev, Kwang-Hee Lee and Chul-Hee Lee
Sensors 2024, 24(14), 4661; https://doi.org/10.3390/s24144661 - 18 Jul 2024
Viewed by 127
Abstract
Deep learning (DL) models require enormous amounts of data to produce reliable diagnosis results. The superiority of DL models over traditional machine learning (ML) methods in terms of feature extraction, feature dimension reduction, and diagnosis performance has been shown in various studies of [...] Read more.
Deep learning (DL) models require enormous amounts of data to produce reliable diagnosis results. The superiority of DL models over traditional machine learning (ML) methods in terms of feature extraction, feature dimension reduction, and diagnosis performance has been shown in various studies of fault diagnosis systems. However, data acquisition can sometimes be compromised by sensor issues, resulting in limited data samples. In this study, we propose a novel DL model based on a stacked convolutional autoencoder (SCAE) to address the challenge of limited data. The innovation of the SCAE model lies in its ability to enhance gradient information flow and extract richer hierarchical features, leading to superior diagnostic performance even with limited and noisy data samples. This article describes the development of a fault diagnosis method for a hydraulic piston pump using time–frequency visual pattern recognition. The proposed SCAE model has been evaluated on limited data samples of a hydraulic piston pump. The findings of the experiment demonstrate that the suggested approach can achieve excellent diagnostic performance with over 99.5% accuracy. Additionally, the SCAE model has outperformed traditional DL models such as deep neural networks (DNN), standard stacked sparse autoencoders (SSAE), and convolutional neural networks (CNN) in terms of diagnosis performance. Furthermore, the proposed model demonstrates robust performance under noisy data conditions, further highlighting its effectiveness and reliability. Full article
(This article belongs to the Special Issue Deep-Learning-Based Defect Detection for Smart Manufacturing)
Show Figures

Figure 1

25 pages, 43361 KiB  
Article
DFFNet: A Rainfall Nowcasting Model Based on Dual-Branch Feature Fusion
by Shuxian Liu, Yulong Liu, Jiong Zheng, Yuanyuan Liao, Guohong Zheng and Yongjun Zhang
Electronics 2024, 13(14), 2826; https://doi.org/10.3390/electronics13142826 - 18 Jul 2024
Viewed by 132
Abstract
Timely and accurate rainfall prediction is crucial to social life and economic activities. Because of the influence of numerous factors on rainfall, making precise predictions is challenging. In this study, the northern Xinjiang region of China is selected as the research area. Based [...] Read more.
Timely and accurate rainfall prediction is crucial to social life and economic activities. Because of the influence of numerous factors on rainfall, making precise predictions is challenging. In this study, the northern Xinjiang region of China is selected as the research area. Based on the pattern of rainfall in the local area and the needs of real life, rainfall is divided into four levels, namely ‘no rain’, ‘light rain’, ‘moderate rain’, and ‘heavy rain and above’, for rainfall levels nowcasting. To solve the problem that the existing model can only extract a single time dependence and cause the loss of some valuable information in rainfall data, a prediction model named DFFNet, which is based on dual-branch feature fusion, is proposed in this paper. The two branches of the model are composed of Transformer and CNN, which are used to extract time dependence and feature interaction in meteorological data, respectively. The features extracted from the two branches are fused for prediction. To verify the performance of DFFNet, the India public rainfall dataset and some sub-datasets in the UEA dataset are chosen for comparison. Compared with the baseline models, DFFNet achieves the best prediction performance on all the selected datasets; compared with the single-branch model, the training time consumption of DFFNet on the two rainfall datasets is reduced by 21% and 9.6%, respectively, and it has a faster convergence speed. The experimental results show that it has certain theoretical value and application value for the study of rainfall nowcasting. Full article
(This article belongs to the Special Issue Application of Big Data Mining and Analysis)
Show Figures

Figure 1

19 pages, 2488 KiB  
Article
Predicting Grapevine Physiological Parameters Using Hyperspectral Remote Sensing Integrated with Hybrid Convolutional Neural Network and Ensemble Stacked Regression
by Prakriti Sharma, Roberto Villegas-Diaz and Anne Fennell
Remote Sens. 2024, 16(14), 2626; https://doi.org/10.3390/rs16142626 - 18 Jul 2024
Viewed by 130
Abstract
Grapevine rootstocks are gaining importance in viticulture as a strategy to combat abiotic challenges, as well as enhance scion physiology. Direct leaf-level physiological parameters like net assimilation rate, stomatal conductance to water vapor, quantum yield of PSII, and transpiration can illuminate the rootstock [...] Read more.
Grapevine rootstocks are gaining importance in viticulture as a strategy to combat abiotic challenges, as well as enhance scion physiology. Direct leaf-level physiological parameters like net assimilation rate, stomatal conductance to water vapor, quantum yield of PSII, and transpiration can illuminate the rootstock effect on scion physiology. However, these measures are time-consuming and limited to leaf-level analysis. This study used different rootstocks to investigate the potential application of aerial hyperspectral imagery in the estimation of canopy level measurements. A statistical framework was developed as an ensemble stacked regression (REGST) that aggregated five different individual machine learning algorithms: Least absolute shrinkage and selection operator (Lasso), Partial least squares regression (PLSR), Ridge regression (RR), Elastic net (ENET), and Principal component regression (PCR) to optimize high-throughput assessment of vine physiology. In addition, a Convolutional Neural Network (CNN) algorithm was integrated into an existing REGST, forming a hybrid CNN-REGST model with the aim of capturing patterns from the hyperspectral signal. Based on the findings, the performance of individual base models exhibited variable prediction accuracies. In most cases, Ridge Regression (RR) demonstrated the lowest test Root Mean Squared Error (RMSE). The ensemble stacked regression model (REGST) outperformed the individual machine learning algorithms with an increase in R2 by (0.03 to 0.1). The performances of CNN-REGST and REGST were similar in estimating the four different traits. Overall, these models were able to explain approximately 55–67% of the variation in the actual ground-truth data. This study suggests that hyperspectral features integrated with powerful AI approaches show great potential in tracing functional traits in grapevines. Full article
Show Figures

Figure 1

19 pages, 2442 KiB  
Article
Prediction of Accident Risk Levels in Traffic Accidents Using Deep Learning and Radial Basis Function Neural Networks Applied to a Dataset with Information on Driving Events
by Cristian Arciniegas-Ayala, Pablo Marcillo, Ángel Leonardo Valdivieso Caraguay and Myriam Hernández-Álvarez
Appl. Sci. 2024, 14(14), 6248; https://doi.org/10.3390/app14146248 - 18 Jul 2024
Viewed by 152
Abstract
A complex AI system must be worked offline because the training and execution phases are processed separately. This process often requires different computer resources due to the high model requirements. A limitation of this approach is the convoluted training process that needs to [...] Read more.
A complex AI system must be worked offline because the training and execution phases are processed separately. This process often requires different computer resources due to the high model requirements. A limitation of this approach is the convoluted training process that needs to be repeated to obtain models with new data continuously incorporated into the knowledge base. Although the environment may be not static, it is crucial to dynamically train models by integrating new information during execution. In this article, artificial neural networks (ANNs) are developed to predict risk levels in traffic accidents with relatively simpler configurations than a deep learning (DL) model, which is more computationally intensive. The objective is to demonstrate that efficient, fast, and comparable results can be obtained using simple architectures such as that offered by the Radial Basis Function neural network (RBFNN). This work led to the generation of a driving dataset, which was subsequently validated for testing ANN models. The driving dataset simulated the dynamic approach by adding new data to the training on-the-fly, given the constant changes in the drivers’ data, vehicle information, environmental conditions, and traffic accidents. This study compares the processing time and performance of a Convolutional Neural Network (CNN), Random Forest (RF), Radial Basis Function (RBF), and Multilayer Perceptron (MLP), using evaluation metrics of accuracy, Specificity, and Sensitivity-recall to recommend an appropriate, simple, and fast ANN architecture that can be implemented in a secure alert traffic system that uses encrypted data. Full article
Show Figures

Figure 1

Back to TopTop