Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,801)

Search Parameters:
Keywords = 1-D convolution

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 853 KiB  
Article
Spatial–Temporal-Correlation-Constrained Dynamic Graph Convolutional Network for Traffic Flow Forecasting
by Yajun Ge, Jiannan Wang, Bo Zhang, Fan Peng, Jing Ma, Chenyu Yang, Yue Zhao and Ming Liu
Mathematics 2024, 12(19), 3159; https://doi.org/10.3390/math12193159 - 9 Oct 2024
Abstract
Accurate traffic flow prediction in road networks is essential for intelligent transportation systems (ITS). Since traffic data are collected from the road network with spatial topological and time series sequences, the traffic flow prediction is regarded as a spatial–temporal prediction task. With the [...] Read more.
Accurate traffic flow prediction in road networks is essential for intelligent transportation systems (ITS). Since traffic data are collected from the road network with spatial topological and time series sequences, the traffic flow prediction is regarded as a spatial–temporal prediction task. With the powerful ability to model the non-Euclidean data, the graph convolutional network (GCN)-based models have become the mainstream framework for traffic forecasting. However, existing GCN-based models either use the manually predefined graph structure to capture the spatial features, ignoring the heterogeneity of road networks, or simply perform 1-D convolution with fixed kernel to capture the temporal dependencies of traffic data, resulting in insufficient long-term temporal feature extraction. To solve those issues, a spatial–temporal correlation constrained dynamic graph convolutional network (STC-DGCN) is proposed for traffic flow forecasting. In STC-DGCN, a spatial–temporal embedding encoder module (STEM) is first constructed to encode the dynamic spatial relationships for road networks at different time steps. Then, a temporal feature encoder module with heterogeneous time series correlation modeling (TFE-HCM) and a spatial feature encoder module with dynamic multi-graph modeling (SFE-DCM) are designed to generate dynamic graph structures for effectively capturing the dynamic spatial and temporal correlations. Finally, a spatial–temporal feature fusion module based on a gating fusion mechanism (STM-GM) is proposed to effectively learn and leverage the inherent spatial–temporal relationships for traffic flow forecasting. Experimental results from three real-world traffic flow datasets demonstrate the superior performance of the proposed STC-DGCN compared with state-of-the-art traffic flow forecasting models. Full article
17 pages, 2983 KiB  
Article
Pose Estimation of a Cobot Implemented on a Small AI-Powered Computing System and a Stereo Camera for Precision Evaluation
by Marco-Antonio Cabrera-Rufino, Juan-Manuel Ramos-Arreguín, Marco-Antonio Aceves-Fernandez, Efren Gorrostieta-Hurtado, Jesus-Carlos Pedraza-Ortega and Juvenal Rodríguez-Resendiz
Biomimetics 2024, 9(10), 610; https://doi.org/10.3390/biomimetics9100610 - 9 Oct 2024
Abstract
The precision of robotic manipulators in the industrial or medical field is very important, especially when it comes to repetitive or exhaustive tasks. Geometric deformations are the most common in this field. For this reason, new robotic vision techniques have been proposed, including [...] Read more.
The precision of robotic manipulators in the industrial or medical field is very important, especially when it comes to repetitive or exhaustive tasks. Geometric deformations are the most common in this field. For this reason, new robotic vision techniques have been proposed, including 3D methods that made it possible to determine the geometric distances between the parts of a robotic manipulator. The aim of this work is to measure the angular position of a robotic arm with six degrees of freedom. For this purpose, a stereo camera and a convolutional neural network algorithm are used to reduce the degradation of precision caused by geometric errors. This method is not intended to replace encoders, but to enhance accuracy by compensating for degradation through an intelligent visual measurement system. The camera is tested and the accuracy is about one millimeter. The implementation of this method leads to better results than traditional and simple neural network methods. Full article
Show Figures

Figure 1

13 pages, 3709 KiB  
Article
An End-to-End Underwater Acoustic Target Recognition Model Based on One-Dimensional Convolution and Transformer
by Kang Yang, Biao Wang, Zide Fang and Banggui Cai
J. Mar. Sci. Eng. 2024, 12(10), 1793; https://doi.org/10.3390/jmse12101793 - 9 Oct 2024
Abstract
Underwater acoustic target recognition (UATR) is crucial for defense and ocean environment monitoring. Although traditional methods and deep learning approaches based on time–frequency domain features have achieved high recognition rates in certain tasks, they rely on manually designed feature extraction processes, leading to [...] Read more.
Underwater acoustic target recognition (UATR) is crucial for defense and ocean environment monitoring. Although traditional methods and deep learning approaches based on time–frequency domain features have achieved high recognition rates in certain tasks, they rely on manually designed feature extraction processes, leading to information loss and limited adaptability to environmental changes. To overcome these limitations, we proposed a novel end-to-end underwater acoustic target recognition model, 1DCTN. This model directly used raw time-domain signals as input, leveraging one-dimensional convolutional neural networks (1D CNNs) to extract local features and combining them with Transformers to capture global dependencies. Our model simplified the recognition process by eliminating the need for complex feature engineering and effectively addressed the limitations of LSTM in handling long-term dependencies. Experimental results on the publicly available ShipsEar dataset demonstrated that 1DCTN achieves a remarkable accuracy of 96.84%, setting a new benchmark for end-to-end models on this dataset. Additionally, 1DCTN stood out among lightweight models, achieving the highest recognition rate, making it a promising direction for future research in underwater acoustic recognition. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

15 pages, 10720 KiB  
Article
Deep-Learning-Based Real-Time Passive Non-Line-of-Sight Imaging for Room-Scale Scenes
by Yuzhe Li and Yuning Zhang
Sensors 2024, 24(19), 6480; https://doi.org/10.3390/s24196480 - 8 Oct 2024
Abstract
Non-line-of-sight imaging is a technique for reconstructing scenes behind obstacles. We report a real-time passive non-line-of-sight (NLOS) imaging method for room-scale hidden scenes, which can be applied to smart home security monitoring sensing systems and indoor fast fuzzy navigation and positioning under the [...] Read more.
Non-line-of-sight imaging is a technique for reconstructing scenes behind obstacles. We report a real-time passive non-line-of-sight (NLOS) imaging method for room-scale hidden scenes, which can be applied to smart home security monitoring sensing systems and indoor fast fuzzy navigation and positioning under the premise of protecting privacy. An unseen scene encoding enhancement network (USEEN) for hidden scene reconstruction is proposed, which is a convolutional neural network designed for NLOS imaging. The network is robust to ambient light interference conditions on diffuse reflective surfaces and maintains a fast reconstruction speed of 12.2 milliseconds per estimation. The consistency of the mean square error (MSE) is verified, and the peak signal-to-noise ratio (PSNR) values of 19.21 dB, 15.86 dB, and 13.62 dB are obtained for the training, validation, and test datasets, respectively. The average values of the structural similarity index (SSIM) are 0.83, 0.68, and 0.59, respectively, and are compared and discussed with the corresponding indicators of the other two models. The sensing system built using this method will show application potential in many fields that require accurate and real-time NLOS imaging, especially smart home security systems in room-scale scenes. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

19 pages, 7895 KiB  
Article
A Novel Trajectory Prediction Method Based on CNN, BiLSTM, and Multi-Head Attention Mechanism
by Yue Xu, Quan Pan, Zengfu Wang and Baoquan Hu
Aerospace 2024, 11(10), 822; https://doi.org/10.3390/aerospace11100822 - 8 Oct 2024
Abstract
A four-dimensional (4D) trajectory is a multi-dimensional time series that embodies rich spatiotemporal features. However, its high complexity and inherent uncertainty pose significant challenges for accurate prediction. In this paper, we present a novel 4D trajectory prediction model that integrates convolutional neural networks [...] Read more.
A four-dimensional (4D) trajectory is a multi-dimensional time series that embodies rich spatiotemporal features. However, its high complexity and inherent uncertainty pose significant challenges for accurate prediction. In this paper, we present a novel 4D trajectory prediction model that integrates convolutional neural networks (CNNs), bidirectional long short-term memory networks (BiLSTMs), and multi-head attention mechanisms. This model effectively addresses the characteristics of aircraft flight trajectories and the difficulties associated with simultaneously extracting spatiotemporal features using existing prediction methods. Specifically, we leverage the local feature extraction capabilities of CNNs to extract key spatial and temporal features from the original trajectory data, such as geometric shape information and dynamic change patterns. The BiLSTM network is employed to consider both forward and backward temporal orders in the trajectory data, allowing for a more comprehensive capture of long-term dependencies. Furthermore, we introduce a multi-head attention mechanism that enhances the model’s ability to accurately identify key information in the trajectory data while minimizing the interference of redundant information. We validated our approach through experiments conducted on a real ADS-B trajectory dataset. The experimental results demonstrate that the proposed method significantly outperforms comparative approaches in terms of trajectory estimation accuracy. Full article
(This article belongs to the Section Aeronautics)
Show Figures

Figure 1

15 pages, 807 KiB  
Article
PointCloud-At: Point Cloud Convolutional Neural Networks with Attention for 3D Data Processing
by Saidu Umar and Aboozar Taherkhani
Sensors 2024, 24(19), 6446; https://doi.org/10.3390/s24196446 - 5 Oct 2024
Abstract
The rapid growth in technologies for 3D sensors has made point cloud data increasingly available in different applications such as autonomous driving, robotics, and virtual and augmented reality. This raises a growing need for deep learning methods to process the data. Point clouds [...] Read more.
The rapid growth in technologies for 3D sensors has made point cloud data increasingly available in different applications such as autonomous driving, robotics, and virtual and augmented reality. This raises a growing need for deep learning methods to process the data. Point clouds are difficult to be used directly as inputs in several deep learning techniques. The difficulty is raised by the unstructured and unordered nature of the point cloud data. So, machine learning models built for images or videos cannot be used directly on point cloud data. Although the research in the field of point clouds has gained high attention and different methods have been developed over the decade, very few research works directly with point cloud data, and most of them convert the point cloud data into 2D images or voxels by performing some pre-processing that causes information loss. Methods that directly work on point clouds are in the early stage and this affects the performance and accuracy of the models. Advanced techniques in classical convolutional neural networks, such as the attention mechanism, need to be transferred to the methods directly working with point clouds. In this research, an attention mechanism is proposed to be added to deep convolutional neural networks that process point clouds directly. The attention module was proposed based on specific pooling operations which are designed to be applied directly to point clouds to extract vital information from the point clouds. Segmentation of the ShapeNet dataset was performed to evaluate the method. The mean intersection over union (mIoU) score of the proposed framework was increased after applying the attention method compared to a base state-of-the-art framework that does not have the attention mechanism. Full article
Show Figures

Figure 1

18 pages, 7988 KiB  
Article
Wind Turbine Bearing Failure Diagnosis Using Multi-Scale Feature Extraction and Residual Neural Networks with Block Attention
by Yuanqing Luo, Yuhang Yang, Shuang Kang, Xueyong Tian, Shiyue Liu and Feng Sun
Actuators 2024, 13(10), 401; https://doi.org/10.3390/act13100401 - 5 Oct 2024
Abstract
Wind turbine rolling bearings are crucial components for ensuring the reliability and stability of wind power systems. Their failure can lead to significant economic losses and equipment downtime. Therefore, the accurate diagnosis of bearing faults is of great importance. Although existing deep learning [...] Read more.
Wind turbine rolling bearings are crucial components for ensuring the reliability and stability of wind power systems. Their failure can lead to significant economic losses and equipment downtime. Therefore, the accurate diagnosis of bearing faults is of great importance. Although existing deep learning fault diagnosis methods have achieved certain results, they still face limitations such as inadequate feature extraction capabilities, insufficient generalization to complex working conditions, and ineffective multi-scale feature capture. To address these issues, this paper proposes an advanced fault diagnosis method named the two-stream feature fusion convolutional neural network (TSFFResNet-Net). Firstly, the proposed method combines the advantages of one-dimensional convolutional neural networks (1D-ResNet) and two-dimensional convolutional neural networks (2D-ResNet). It transforms one-dimensional vibration signals into two-dimensional images through the empirical wavelet transform (EWT) method. Then, parallel convolutional kernels in 1D-ResNet and 2D-ResNet are used to extract multi-scale features, respectively. Next, the Convolutional Block Attention Module (CBAM) is introduced to enhance the network’s ability to capture key features by focusing on important features in specific channels or spatial areas. After feature fusion, CBAM is introduced again to further enhance the effect of feature fusion, ensuring that the features extracted by different network branches can be effectively integrated, ultimately providing more accurate input features for the classification task of the fully connected layer. The experimental results demonstrate that the proposed method outperforms other traditional methods and advanced convolutional neural network models on different datasets. Compared with convolutional neural network models such as LeNet-5, AlexNet, and ResNet, the proposed method achieves a significantly higher accuracy on the test set, with a stable accuracy of over 99%. Compared with other models, it shows better generalization and stability, effectively improving the overall performance of rolling bearing vibration signal fault diagnosis. The method provides an effective solution for the intelligent fault diagnosis of wind turbine rolling bearings. Full article
(This article belongs to the Section Actuators for Manufacturing Systems)
Show Figures

Figure 1

12 pages, 1419 KiB  
Article
Mobile Spatiotemporal Gait Segmentation Using an Ear-Worn Motion Sensor and Deep Learning
by Julian Decker, Lukas Boborzi, Roman Schniepp, Klaus Jahn and Max Wuehr
Sensors 2024, 24(19), 6442; https://doi.org/10.3390/s24196442 - 4 Oct 2024
Abstract
Mobile health technologies enable continuous, quantitative assessment of mobility and gait in real-world environments, facilitating early diagnoses of gait disorders, disease progression monitoring, and prediction of adverse events like falls. Traditionally, mobile gait assessment predominantly relied on body-fixed sensors positioned at the feet [...] Read more.
Mobile health technologies enable continuous, quantitative assessment of mobility and gait in real-world environments, facilitating early diagnoses of gait disorders, disease progression monitoring, and prediction of adverse events like falls. Traditionally, mobile gait assessment predominantly relied on body-fixed sensors positioned at the feet or lower trunk. Here, we investigate the potential of an algorithm utilizing an ear-worn motion sensor for spatiotemporal segmentation of gait patterns. We collected 3D acceleration profiles from the ear-worn sensor during varied walking speeds in 53 healthy adults. Temporal convolutional networks were trained to detect stepping sequences and predict spatial relations between steps. The resulting algorithm, mEar, accurately detects initial and final ground contacts (F1 score of 99% and 91%, respectively). It enables the determination of temporal and spatial gait cycle characteristics (among others, stride time and stride length) with good to excellent validity at a precision sufficient to monitor clinically relevant changes in walking speed, stride-to-stride variability, and side asymmetry. This study highlights the ear as a viable site for monitoring gait and proposes its potential integration with in-ear vital-sign monitoring. Such integration offers a practical approach to comprehensive health monitoring and telemedical applications, by integrating multiple sensors in a single anatomical location. Full article
(This article belongs to the Special Issue Sensors and Artificial Intelligence in Gait and Posture Analysis)
Show Figures

Figure 1

23 pages, 1046 KiB  
Article
Forecasting Short- and Long-Term Wind Speed in Limpopo Province Using Machine Learning and Extreme Value Theory
by Kgothatso Makubyane and Daniel Maposa
Forecasting 2024, 6(4), 885-907; https://doi.org/10.3390/forecast6040044 - 4 Oct 2024
Abstract
This study investigates wind speed prediction using advanced machine learning techniques, comparing the performance of Vanilla long short-term memory (LSTM) and convolutional neural network (CNN) models, alongside the application of extreme value theory (EVT) using the r-largest order generalised extreme value distribution ( [...] Read more.
This study investigates wind speed prediction using advanced machine learning techniques, comparing the performance of Vanilla long short-term memory (LSTM) and convolutional neural network (CNN) models, alongside the application of extreme value theory (EVT) using the r-largest order generalised extreme value distribution (GEVDr). Over the past couple of decades, the academic literature has transitioned from conventional statistical time series models to embracing EVT and machine learning algorithms for the modelling of environmental variables. This study adds value to the literature and knowledge of modelling wind speed using both EVT and machine learning. The primary aim of this study is to forecast wind speed in the Limpopo province of South Africa to showcase the dependability and potential of wind power generation. The application of CNN showcased considerable predictive accuracy compared to the Vanilla LSTM, achieving 88.66% accuracy with monthly time steps. The CNN predictions for the next five years, in m/s, were 9.91 (2024), 7.64 (2025), 7.81 (2026), 7.13 (2027), and 9.59 (2028), slightly outperforming the Vanilla LSTM, which predicted 9.43 (2024), 7.75 (2025), 7.85 (2026), 6.87 (2027), and 9.43 (2028). This highlights CNN’s superior ability to capture complex patterns in wind speed dynamics over time. Concurrently, the analysis of the GEVDr across various order statistics identified GEVDr=2 as the optimal model, supported by its favourable evaluation metrics in terms of Akaike information criteria (AIC) and Bayesian information criteria (BIC). The 300-year return level for GEVDr=2 was found to be 22.89 m/s, indicating a rare wind speed event. Seasonal wind speed analysis revealed distinct patterns, with winter emerging as the most efficient season for wind, featuring a median wind speed of 7.96 m/s. Future research could focus on enhancing prediction accuracy through hybrid algorithms and incorporating additional meteorological variables. To the best of our knowledge, this is the first study to successfully combine EVT and machine learning for short- and long-term wind speed forecasting, providing a novel framework for reliable wind energy planning. Full article
Show Figures

Figure 1

13 pages, 1856 KiB  
Article
Assessment of the Breast Density Prevalence in Swiss Women with a Deep Convolutional Neural Network: A Cross-Sectional Study
by Adergicia V. Kaiser, Daniela Zanolin-Purin, Natalie Chuck, Jennifer Enaux and Daniela Wruk
Diagnostics 2024, 14(19), 2212; https://doi.org/10.3390/diagnostics14192212 - 3 Oct 2024
Abstract
Background/Objectives: High breast density is a risk factor for breast cancer and can reduce the sensitivity of mammography. Given the influence of breast density on patient risk stratification and screening accuracy, it is crucial to monitor the prevalence of extremely dense breasts within [...] Read more.
Background/Objectives: High breast density is a risk factor for breast cancer and can reduce the sensitivity of mammography. Given the influence of breast density on patient risk stratification and screening accuracy, it is crucial to monitor the prevalence of extremely dense breasts within local populations. Moreover, there is a lack of comprehensive understanding regarding breast density prevalence in Switzerland. Therefore, this study aimed to determine the prevalence of breast density in a selected Swiss population. Methods: To overcome the potential variability in breast density classifications by human readers, this study utilized commercially available deep convolutional neural network breast classification software. A retrospective analysis of mammographic images of women aged 40 years and older was performed. Results: A total of 4698 mammograms from women (58 ± 11 years) were included in this study. The highest prevalence of breast density was in category C (heterogeneously dense), which was observed in 41.5% of the cases. This was followed by category B (scattered areas of fibroglandular tissue), which accounted for 22.5%. Conclusions: Notably, extremely dense breasts (category D) were significantly more common in younger women, with a prevalence of 34%. However, this rate dropped sharply to less than 10% in women over 55 years of age. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

16 pages, 4191 KiB  
Article
Respiratory Rate Estimation from Thermal Video Data Using Spatio-Temporal Deep Learning
by Mohsen Mozafari, Andrew J. Law, Rafik A. Goubran and James R. Green
Sensors 2024, 24(19), 6386; https://doi.org/10.3390/s24196386 - 2 Oct 2024
Abstract
Thermal videos provide a privacy-preserving yet information-rich data source for remote health monitoring, especially for respiration rate (RR) estimation. This paper introduces an end-to-end deep learning approach to RR measurement using thermal video data. A detection transformer (DeTr) first finds the subject’s facial [...] Read more.
Thermal videos provide a privacy-preserving yet information-rich data source for remote health monitoring, especially for respiration rate (RR) estimation. This paper introduces an end-to-end deep learning approach to RR measurement using thermal video data. A detection transformer (DeTr) first finds the subject’s facial region of interest in each thermal frame. A respiratory signal is estimated from a dynamically cropped thermal video using 3D convolutional neural networks and bi-directional long short-term memory stages. To account for the expected phase shift between the respiration measured using a respiratory effort belt vs. a facial video, a novel loss function based on negative maximum cross-correlation and absolute frequency peak difference was introduced. Thermal recordings from 22 subjects, with simultaneous gold standard respiratory effort measurements, were studied while sitting or standing, both with and without a face mask. The RR estimation results showed that our proposed method outperformed existing models, achieving an error of only 1.6 breaths per minute across the four conditions. The proposed method sets a new State-of-the-Art for RR estimation accuracy, while still permitting real-time RR estimation. Full article
(This article belongs to the Special Issue Machine Learning and Image-Based Smart Sensing and Applications)
Show Figures

Figure 1

20 pages, 3901 KiB  
Article
Multi-Modal Fusion Network with Multi-Head Self-Attention for Injection Training Evaluation in Medical Education
by Zhe Li, Aya Kanazuka, Atsushi Hojo, Yukihiro Nomura and Toshiya Nakaguchi
Electronics 2024, 13(19), 3882; https://doi.org/10.3390/electronics13193882 - 30 Sep 2024
Abstract
The COVID-19 pandemic has significantly disrupted traditional medical training, particularly in critical areas such as the injection process, which require expert supervision. To address the challenges posed by reduced face-to-face interactions, this study introduces a multi-modal fusion network designed to evaluate the timing [...] Read more.
The COVID-19 pandemic has significantly disrupted traditional medical training, particularly in critical areas such as the injection process, which require expert supervision. To address the challenges posed by reduced face-to-face interactions, this study introduces a multi-modal fusion network designed to evaluate the timing and motion aspects of the injection training process in medical education. The proposed framework integrates 3D reconstructed data and 2D images of hand movements during the injection process. The 3D data are preprocessed and encoded by a Long Short-Term Memory (LSTM) network to extract temporal features, while a Convolutional Neural Network (CNN) processes the 2D images to capture detailed image features. These encoded features are then fused and refined through a proposed multi-head self-attention module, which enhances the model’s ability to capture and weigh important temporal and image dynamics in the injection process. The final classification of the injection process is conducted by a classifier module. The model’s performance was rigorously evaluated using video data from 255 subjects with assessments made by professional physicians according to the Objective Structured Assessment of Technical Skill—Global Rating Score (OSATS-GRS)[B] criteria for time and motion evaluation. The experimental results demonstrate that the proposed data fusion model achieves an accuracy of 0.7238, an F1-score of 0.7060, a precision of 0.7339, a recall of 0.7238, and an AUC of 0.8343. These findings highlight the model’s potential as an effective tool for providing objective feedback in medical injection training, offering a scalable solution for the post-pandemic evolution of medical education. Full article
Show Figures

Figure 1

22 pages, 5120 KiB  
Article
A Representation-Learning-Based Graph and Generative Network for Hyperspectral Small Target Detection
by Yunsong Li, Jiaping Zhong, Weiying Xie and Paolo Gamba
Remote Sens. 2024, 16(19), 3638; https://doi.org/10.3390/rs16193638 - 29 Sep 2024
Abstract
Hyperspectral small target detection (HSTD) is a promising pixel-level detection task. However, due to the low contrast and imbalanced number between the target and the background spatially and the high dimensions spectrally, it is a challenging one. To address these issues, this work [...] Read more.
Hyperspectral small target detection (HSTD) is a promising pixel-level detection task. However, due to the low contrast and imbalanced number between the target and the background spatially and the high dimensions spectrally, it is a challenging one. To address these issues, this work proposes a representation-learning-based graph and generative network for hyperspectral small target detection. The model builds a fusion network through frequency representation for HSTD, where the novel architecture incorporates irregular topological data and spatial–spectral features to improve its representation ability. Firstly, a Graph Convolutional Network (GCN) module better models the non-local topological relationship between samples to represent the hyperspectral scene’s underlying data structure. The mini-batch-training pattern of the GCN decreases the high computational cost of building an adjacency matrix for high-dimensional data sets. In parallel, the generative model enhances the differentiation reconstruction and the deep feature representation ability with respect to the target spectral signature. Finally, a fusion module compensates for the extracted different types of HS features and integrates their complementary merits for hyperspectral data interpretation while increasing the detection and background suppression capabilities. The performance of the proposed approach is evaluated using the average scores of AUCD,F, AUCF,τ, AUCBS, and AUCSNPR. The corresponding values are 0.99660, 0.00078, 0.99587, and 333.629, respectively. These results demonstrate the accuracy of the model in different evaluation metrics, with AUCD,F achieving the highest score, indicating strong detection performance across varying thresholds. Experiments on different hyperspectral data sets demonstrate the advantages of the proposed architecture. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

19 pages, 7867 KiB  
Article
Advanced 3D Face Reconstruction from Single 2D Images Using Enhanced Adversarial Neural Networks and Graph Neural Networks
by Mohamed Fathallah, Sherif Eletriby, Maazen Alsabaan, Mohamed I. Ibrahem and Gamal Farok
Sensors 2024, 24(19), 6280; https://doi.org/10.3390/s24196280 - 28 Sep 2024
Abstract
This paper presents a novel framework for 3D face reconstruction from single 2D images and addresses critical limitations in existing methods. Our approach integrates modified adversarial neural networks with graph neural networks to achieve state-of-the-art performance. Key innovations include (1) a generator architecture [...] Read more.
This paper presents a novel framework for 3D face reconstruction from single 2D images and addresses critical limitations in existing methods. Our approach integrates modified adversarial neural networks with graph neural networks to achieve state-of-the-art performance. Key innovations include (1) a generator architecture based on Graph Convolutional Networks (GCNs) with a novel loss function and identity blocks, mitigating mode collapse and instability; (2) the integration of facial landmarks and a non-parametric efficient-net decoder for enhanced feature capture; and (3) a lightweight GCN-based discriminator for improved accuracy and stability. Evaluated on the 300W-LP and AFLW2000-3D datasets, our method outperforms existing approaches, reducing Chamfer Distance by 62.7% and Earth Mover’s Distance by 57.1% on 300W-LP. Moreover, our framework demonstrates superior robustness to variations in head positioning, occlusion, noise, and lighting conditions while achieving significantly faster processing times. Full article
(This article belongs to the Collection 3D Imaging and Sensing System)
Show Figures

Figure 1

20 pages, 14487 KiB  
Article
Fault Classification of 3D-Printing Operations Using Different Types of Machine and Deep Learning Techniques
by Satish Kumar, Sameer Sayyad and Arunkumar Bongale
AI 2024, 5(4), 1759-1778; https://doi.org/10.3390/ai5040087 - 27 Sep 2024
Abstract
Fused deposition modeling (FDM), a method of additive manufacturing (AM), comprises the extrusion of materials via a nozzle and the subsequent combining of the layers to create 3D-printed objects. FDM is a widely used method for 3D-printing objects since it is affordable, effective, [...] Read more.
Fused deposition modeling (FDM), a method of additive manufacturing (AM), comprises the extrusion of materials via a nozzle and the subsequent combining of the layers to create 3D-printed objects. FDM is a widely used method for 3D-printing objects since it is affordable, effective, and easy to use. Some defects such as poor infill, elephant foot, layer shift, and poor surface finish arise in the FDM components at the printing stage due to variations in printing parameters such as printing speed, change in nozzle, or bed temperature. Proper fault classification is required to identify the cause of faulty products. In this work, the multi-sensory data are gathered using different sensors such as vibration, current, temperature, and sound sensors. The data acquisition is performed by using the National Instrumentation (NI) Data Acquisition System (DAQ) which provides the synchronous multi-sensory data for the model training. To induce the faults, the data are captured under different conditions such as variations in printing speed, temperate, and jerk during the printing. The collected data are used to train the machine learning (ML) and deep learning (DL) classification models to classify the variation in printing parameters. The ML models such as k-nearest neighbor (KNN), decision tree (DT), extra trees (ET), and random forest (RF) with convolutional neural network (CNN) as a DL model are used to classify the variable operation printing parameters. Out of the available models, in ML models, the RF classifier shows a classification accuracy of around 91% whereas, in the DL model, the CNN model shows good classification performance with accuracy ranging from 92 to 94% under variable operating conditions. Full article
(This article belongs to the Special Issue Intelligent Systems for Industry 4.0)
Show Figures

Figure 1

Back to TopTop