Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (9,000)

Search Parameters:
Keywords = deep convolutional neural networks

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 3824 KiB  
Article
A Hybrid Network Integrating MHSA and 1D CNN–Bi-LSTM for Interference Mitigation in Faster-than-Nyquist MIMO Optical Wireless Communications
by Minghua Cao, Qing Yang, Genxue Zhou, Yue Zhang, Xia Zhang and Huiqin Wang
Photonics 2024, 11(10), 982; https://doi.org/10.3390/photonics11100982 (registering DOI) - 19 Oct 2024
Viewed by 111
Abstract
To mitigate inter-symbol interference (ISI) caused by Faster-than-Nyquist (FTN) technology in a multiple input multiple output (MIMO) optical wireless communication (OWC) system, we propose an ISI cancellation algorithm that combines multi-head self-attention (MHSA), a one-dimensional convolutional neural network (1D CNN), and bi-directional long [...] Read more.
To mitigate inter-symbol interference (ISI) caused by Faster-than-Nyquist (FTN) technology in a multiple input multiple output (MIMO) optical wireless communication (OWC) system, we propose an ISI cancellation algorithm that combines multi-head self-attention (MHSA), a one-dimensional convolutional neural network (1D CNN), and bi-directional long short-term memory (Bi-LSTM). This hybrid network extracts data features using 1D CNN and captures sequential information with Bi-LSTM, while incorporating MHSA to comprehensively reduce ISI. We analyze the impact of antenna numbers, acceleration factors, wavelength, and turbulence intensity on the system’s bit error rate (BER) performance. Additionally, we compare the waveform graphs and amplitude–frequency characteristics of FTN signals before and after processing, specifically comparing sampled values of four-pulse-amplitude modulation (4PAM) signals with those obtained after ISI cancellation. The simulation results demonstrate that within the Mazo limit for selecting acceleration factors, our proposal achieves a 7 dB improvement in BER compared to the conventional systems without deep learning (DL)-based ISI cancellation algorithms. Furthermore, compared to systems employing a point-by-point elimination adaptive pre-equalization algorithm, our proposal exhibits comparable BER performance to orthogonal transmission systems while reducing computational complexity by 31.15%. Full article
(This article belongs to the Special Issue Advanced Technologies in Optical Wireless Communications)
Show Figures

Figure 1

17 pages, 5181 KiB  
Article
Shift-Invariance Robustness of Convolutional Neural Networks in Side-Channel Analysis
by Marina Krček, Lichao Wu, Guilherme Perin and Stjepan Picek
Mathematics 2024, 12(20), 3279; https://doi.org/10.3390/math12203279 (registering DOI) - 18 Oct 2024
Viewed by 262
Abstract
Convolutional neural networks (CNNs) offer unrivaled performance in profiling side-channel analysis. This claim is corroborated by numerous results where CNNs break targets protected with masking and hiding countermeasures. One hiding countermeasure commonly investigated in related works is desynchronization (misalignment). The conclusions usually state [...] Read more.
Convolutional neural networks (CNNs) offer unrivaled performance in profiling side-channel analysis. This claim is corroborated by numerous results where CNNs break targets protected with masking and hiding countermeasures. One hiding countermeasure commonly investigated in related works is desynchronization (misalignment). The conclusions usually state that CNNs can break desynchronization as they are shift-invariant. This paper investigates that claim in more detail and reveals that the situation is more complex. While CNNs have certain shift-invariance, it is insufficient for commonly encountered scenarios in deep learning-based side-channel analysis. We investigate data augmentation to improve the shift-invariance and, in a more powerful version, ensembles of data augmentation. Our results show that the proposed techniques work very well and improve the attack significantly, even for an order of magnitude. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence to Cryptography)
Show Figures

Figure 1

20 pages, 1638 KiB  
Article
GPT-Driven Radiology Report Generation with Fine-Tuned Llama 3
by Ștefan-Vlad Voinea, Mădălin Mămuleanu, Rossy Vlăduț Teică, Lucian Mihai Florescu, Dan Selișteanu and Ioana Andreea Gheonea
Bioengineering 2024, 11(10), 1043; https://doi.org/10.3390/bioengineering11101043 (registering DOI) - 18 Oct 2024
Viewed by 144
Abstract
The integration of deep learning into radiology has the potential to enhance diagnostic processes, yet its acceptance in clinical practice remains limited due to various challenges. This study aimed to develop and evaluate a fine-tuned large language model (LLM), based on Llama 3-8B, [...] Read more.
The integration of deep learning into radiology has the potential to enhance diagnostic processes, yet its acceptance in clinical practice remains limited due to various challenges. This study aimed to develop and evaluate a fine-tuned large language model (LLM), based on Llama 3-8B, to automate the generation of accurate and concise conclusions in magnetic resonance imaging (MRI) and computed tomography (CT) radiology reports, thereby assisting radiologists and improving reporting efficiency. A dataset comprising 15,000 radiology reports was collected from the University of Medicine and Pharmacy of Craiova’s Imaging Center, covering a diverse range of MRI and CT examinations made by four experienced radiologists. The Llama 3-8B model was fine-tuned using transfer-learning techniques, incorporating parameter quantization to 4-bit precision and low-rank adaptation (LoRA) with a rank of 16 to optimize computational efficiency on consumer-grade GPUs. The model was trained over five epochs using an NVIDIA RTX 3090 GPU, with intermediary checkpoints saved for monitoring. Performance was evaluated quantitatively using Bidirectional Encoder Representations from Transformers Score (BERTScore), Recall-Oriented Understudy for Gisting Evaluation (ROUGE), Bilingual Evaluation Understudy (BLEU), and Metric for Evaluation of Translation with Explicit Ordering (METEOR) metrics on a held-out test set. Additionally, a qualitative assessment was conducted, involving 13 independent radiologists who participated in a Turing-like test and provided ratings for the AI-generated conclusions. The fine-tuned model demonstrated strong quantitative performance, achieving a BERTScore F1 of 0.8054, a ROUGE-1 F1 of 0.4998, a ROUGE-L F1 of 0.4628, and a METEOR score of 0.4282. In the human evaluation, the artificial intelligence (AI)-generated conclusions were preferred over human-written ones in approximately 21.8% of cases, indicating that the model’s outputs were competitive with those of experienced radiologists. The average rating of the AI-generated conclusions was 3.65 out of 5, reflecting a generally favorable assessment. Notably, the model maintained its consistency across various types of reports and demonstrated the ability to generalize to unseen data. The fine-tuned Llama 3-8B model effectively generates accurate and coherent conclusions for MRI and CT radiology reports. By automating the conclusion-writing process, this approach can assist radiologists in reducing their workload and enhancing report consistency, potentially addressing some barriers to the adoption of deep learning in clinical practice. The positive evaluations from independent radiologists underscore the model’s potential utility. While the model demonstrated strong performance, limitations such as dataset bias, limited sample diversity, a lack of clinical judgment, and the need for large computational resources require further refinement and real-world validation. Future work should explore the integration of such models into clinical workflows, address ethical and legal considerations, and extend this approach to generate complete radiology reports. Full article
17 pages, 1085 KiB  
Article
Enhancing Brain Tumor Diagnosis with L-Net: A Novel Deep Learning Approach for MRI Image Segmentation and Classification
by Lehel Dénes-Fazakas, Levente Kovács, György Eigner and László Szilágyi
Biomedicines 2024, 12(10), 2388; https://doi.org/10.3390/biomedicines12102388 (registering DOI) - 18 Oct 2024
Viewed by 173
Abstract
Background: Brain tumors are highly complex, making their detection and classification a significant challenge in modern medical diagnostics. The accurate segmentation and classification of brain tumors from MRI images are crucial for effective treatment planning. This study aims to develop an advanced neural [...] Read more.
Background: Brain tumors are highly complex, making their detection and classification a significant challenge in modern medical diagnostics. The accurate segmentation and classification of brain tumors from MRI images are crucial for effective treatment planning. This study aims to develop an advanced neural network architecture that addresses these challenges. Methods: We propose L-net, a novel architecture combining U-net for tumor boundary segmentation and a convolutional neural network (CNN) for tumor classification. These two units are coupled such a way that the CNN classifies the MRI images based on the features extracted by the U-net while segmenting the tumor, instead of relying on the original input images. The model is trained on a dataset of 3064 high-resolution MRI images, encompassing gliomas, meningiomas, and pituitary tumors, ensuring robust performance across different tumor types. Results: L-net achieved a classification accuracy of up to 99.6%, surpassing existing models in both segmentation and classification tasks. The model demonstrated effectiveness even with lower image resolutions, making it suitable for diverse clinical settings. Conclusions: The proposed L-net model provides an accurate and unified approach to brain tumor segmentation and classification. Its enhanced performance contributes to more reliable and precise diagnosis, supporting early detection and treatment in clinical applications. Full article
(This article belongs to the Special Issue Diagnosis, Pathogenesis and Treatment of CNS Tumors)
16 pages, 2542 KiB  
Article
Deep Learning-Based Reconstruction of 3D Morphology of Geomaterial Particles from Single-View 2D Images
by Jiangpeng Zhao, Heping Xie, Cunbao Li and Yifei Liu
Materials 2024, 17(20), 5100; https://doi.org/10.3390/ma17205100 (registering DOI) - 18 Oct 2024
Viewed by 137
Abstract
The morphology of particles formed in different environments contains critical information. Thus, the rapid and effective reconstruction of their three-dimensional (3D) morphology is crucial. This study reconstructs the 3D morphology from two-dimensional (2D) images of particles using artificial intelligence (AI). More than 100,000 [...] Read more.
The morphology of particles formed in different environments contains critical information. Thus, the rapid and effective reconstruction of their three-dimensional (3D) morphology is crucial. This study reconstructs the 3D morphology from two-dimensional (2D) images of particles using artificial intelligence (AI). More than 100,000 particles were sampled from three sources: naturally formed particles (desert sand), manufactured particles (lunar soil simulant), and numerically generated digital particles. A deep learning approach based on a voxel representation of the morphology and multi-dimensional convolutional neural networks was proposed to rapidly upscale and reconstruct particle morphology. The trained model was tested using the three particle types and evaluated using different multi-scale morphological descriptors. The results demonstrated that the statistical properties of the morphological descriptors were consistent for the real 3D particles and those derived from the 2D images and the model. This finding confirms the model’s validity and generalizability in upscaling and reconstructing diverse particle samples. This study provides a method for generating 3D numerical representations of geological particles, facilitating in-depth analysis of properties, such as mechanical behavior and transport characteristics, from 2D images. Full article
25 pages, 3905 KiB  
Article
An Efficient Ship Detection Method Based on YOLO and Ship Wakes Using High-Resolution Optical Jilin1 Satellite Imagery
by Fangli Mou, Zide Fan, Yunping Ge, Lei Wang and Xinming Li
Sensors 2024, 24(20), 6708; https://doi.org/10.3390/s24206708 (registering DOI) - 18 Oct 2024
Viewed by 113
Abstract
In this study, we propose a practical and efficient scheme for ship detection in remote sensing imagery. Our method is developed using both ship body detection and ship wake detection and combines deep learning and feature-based image processing. A deep convolutional neural network [...] Read more.
In this study, we propose a practical and efficient scheme for ship detection in remote sensing imagery. Our method is developed using both ship body detection and ship wake detection and combines deep learning and feature-based image processing. A deep convolutional neural network is used to achieve ship body detection, and a feature-based processing method is proposed to detect ship wakes. For better analysis, we model the sea region and evaluate the quality of the image. Generally, the wake detection result is used to assist ship detection and obtain the sailing direction. Conventional methods cannot detect ships that are covered by clouds or outside the image boundary. The method proposed in this paper uses the wake to detect such ships, with a certain level of confidence and low false alarm probability in detection. Practical aspects such as the method’s applicability and time efficiency are considered in our method for real applications. We demonstrate the effectiveness of our method in a real remote sensing dataset. The results show that over 93.5% of ships and over 70% of targets with no visible ship body can be successfully detected. This illustrates that the proposed detection framework can fill the gap regarding the detection of sailing ships in a remote sensing image. Full article
(This article belongs to the Section Remote Sensors)
20 pages, 10555 KiB  
Article
Cloud Detection Using a UNet3+ Model with a Hybrid Swin Transformer and EfficientNet (UNet3+STE) for Very-High-Resolution Satellite Imagery
by Jaewan Choi, Doochun Seo, Jinha Jung, Youkyung Han, Jaehong Oh and Changno Lee
Remote Sens. 2024, 16(20), 3880; https://doi.org/10.3390/rs16203880 (registering DOI) - 18 Oct 2024
Viewed by 190
Abstract
It is necessary to extract and recognize the cloud regions presented in imagery to generate satellite imagery as analysis-ready data (ARD). In this manuscript, we proposed a new deep learning model to detect cloud areas in very-high-resolution (VHR) satellite imagery by fusing two [...] Read more.
It is necessary to extract and recognize the cloud regions presented in imagery to generate satellite imagery as analysis-ready data (ARD). In this manuscript, we proposed a new deep learning model to detect cloud areas in very-high-resolution (VHR) satellite imagery by fusing two deep learning architectures. The proposed UNet3+ model with a hybrid Swin Transformer and EfficientNet (UNet3+STE) was based on the structure of UNet3+, with the encoder sequentially combining EfficientNet based on mobile inverted bottleneck convolution (MBConv) and the Swin Transformer. By sequentially utilizing convolutional neural networks (CNNs) and transformer layers, the proposed algorithm aimed to extract the local and global information of cloud regions effectively. In addition, the decoder used MBConv to restore the spatial information of the feature map extracted by the encoder and adopted the deep supervision strategy of UNet3+ to enhance the model’s performance. The proposed model was trained using the open dataset derived from KOMPSAT-3 and 3A satellite imagery and conducted a comparative evaluation with the state-of-the-art (SOTA) methods on fourteen test datasets at the product level. The experimental results confirmed that the proposed UNet3+STE model outperformed the SOTA methods and demonstrated the most stable precision, recall, and F1 score values with fewer parameters and lower complexity. Full article
(This article belongs to the Special Issue Deep Learning Techniques Applied in Remote Sensing)
Show Figures

Figure 1

22 pages, 16009 KiB  
Article
Lightweight Multi-Domain Fusion Model for Through-Wall Human Activity Recognition Using IR-UWB Radar
by Ling Huang, Dong Lei, Bowen Zheng, Guiping Chen, Huifeng An and Mingxuan Li
Appl. Sci. 2024, 14(20), 9522; https://doi.org/10.3390/app14209522 - 18 Oct 2024
Viewed by 251
Abstract
Impulse radio ultra-wideband (IR-UWB) radar, operating in the low-frequency band, can penetrate walls and utilize its high range resolution to recognize different human activities. Complex deep neural networks have demonstrated significant performance advantages in classifying radar spectrograms of various actions, but at the [...] Read more.
Impulse radio ultra-wideband (IR-UWB) radar, operating in the low-frequency band, can penetrate walls and utilize its high range resolution to recognize different human activities. Complex deep neural networks have demonstrated significant performance advantages in classifying radar spectrograms of various actions, but at the cost of a substantial computational overhead. In response, this paper proposes a lightweight model named TG2-CAFNet. First, clutter suppression and time–frequency analysis are used to obtain range–time and micro-Doppler feature maps of human activities. Then, leveraging GhostV2 convolution, a lightweight feature extraction module, TG2, suitable for radar spectrograms is constructed. Using a parallel structure, the features of the two spectrograms are extracted separately. Finally, to further explore the correlation between the two spectrograms and enhance the feature representation capabilities, an improved nonlinear fusion method called coordinate attention fusion (CAF) is proposed based on attention feature fusion (AFF). This method extends the adaptive weighting fusion of AFF to a spatial distribution, effectively capturing the subtle spatial relationships between the two radar spectrograms. Experiments showed that the proposed method achieved a high degree of model lightweightness, while also achieving a recognition accuracy of 99.1%. Full article
Show Figures

Figure 1

30 pages, 39939 KiB  
Article
Urban Color Perception and Sentiment Analysis Based on Deep Learning and Street View Big Data
by Mingyang Yu, Xiangyu Zheng, Pinrui Qin, Weikang Cui and Qingrui Ji
Appl. Sci. 2024, 14(20), 9521; https://doi.org/10.3390/app14209521 - 18 Oct 2024
Viewed by 275
Abstract
The acceleration of urbanization has resulted in a heightened awareness of the impacts of urban environments on residents’ emotional states. This present study focuses on the Lixia District of Jinan City. By using urban street view big data and deep learning methods, we [...] Read more.
The acceleration of urbanization has resulted in a heightened awareness of the impacts of urban environments on residents’ emotional states. This present study focuses on the Lixia District of Jinan City. By using urban street view big data and deep learning methods, we undertook a detailed analysis of the impacts of urban color features on residents’ emotional perceptions. In particular, a substantial corpus of street scene image data was extracted and processed. This was performed using a deep convolutional neural network (DCNN) and semantic segmentation technology (PSPNet), which enabled the simulation and prediction of the subjective perception of the urban environment by humans. Furthermore, the color complexity and coordination in the street scene were quantified and combined with residents’ emotional feedback to carry out a multi-dimensional analysis. The findings revealed that color complexity and coordination were significant elements influencing residents’ emotional perceptions. A high color complexity is visually appealing, but can lead to fatigue, discomfort, and boredom; a moderate complexity stimulates vitality and pleasure; high levels of regional harmony and aesthetics can increase perceptions of beauty and security; and low levels of coordination can increase feelings of depression. The environmental characteristics of different areas and differences in the daily activities of residents resulted in regional differences regarding the impacts of color features on emotional perception. This study corroborates the assertion that environmental color coordination has the capacity to enhance residents’ emotions, thereby providing an important reference point for urban planning. Planning should be based on the functional characteristics of the region, and color complexity and coordination should be reasonably regulated to optimize the emotional experiences of residents. Differentiated color management enhances urban aesthetics, livability, and residents’ happiness and promotes sustainable development. In the future, the influences of color and environmental factors on emotions can be explored in depth, with a view to assist in the formulation of fine urban design. Full article
Show Figures

Figure 1

14 pages, 3525 KiB  
Article
Deep Learning-Based Gender Recognition in Cherry Valley Ducks Through Sound Analysis
by Guofeng Han, Yujing Liu, Jiawen Cai, Enze Duan, Zefeng Shi, Shida Zhao, Lianfei Huo, Huixin Wang and Zongchun Bai
Animals 2024, 14(20), 3017; https://doi.org/10.3390/ani14203017 - 18 Oct 2024
Viewed by 157
Abstract
Gender recognition is an important part of the duck industry. Currently, the gender identification of ducks mainly relies on manual labor, which is highly labor-intensive. This study aims to propose a novel method for distinguishing between males and females based on the characteristic [...] Read more.
Gender recognition is an important part of the duck industry. Currently, the gender identification of ducks mainly relies on manual labor, which is highly labor-intensive. This study aims to propose a novel method for distinguishing between males and females based on the characteristic sound parameters for day-old ducks. The effective data from the sounds of day-old ducks were recorded and extracted using the endpoint detection method. The 12-dimensional Mel-frequency cepstral coefficients (MFCCs) with first-order and second-order difference coefficients in the effective sound signals of the ducks were calculated, and a total of 36-dimensional feature vectors were obtained. These data were used as input information to train three classification models, include a backpropagation neural network (BPNN), a deep neural network (DNN), and a convolutional neural network (CNN). The training results show that the accuracies of the BPNN, DNN, and CNN were 83.87%, 83.94%, and 84.15%, respectively, and that the three classification models could identify the sounds of male and female ducks. The prediction results showed that the prediction accuracies of the BPNN, DNN, and CNN were 93.33%, 91.67%, and 95.0%, respectively, which shows that the scheme for distinguishing between male and female ducks via sound had high accuracy. Moreover, the CNN demonstrated the best recognition effect. The method proposed in this study can provide some support for developing an efficient technique for gender identification in duck production. Full article
Show Figures

Figure 1

8 pages, 1124 KiB  
Proceeding Paper
A Fog Computing-Based Cost-Effective Smart Health Monitoring Device for Infectious Disease Applications
by Saranya Govindakumar, Vijayalakshmi Sankaran, Paramasivam Alagumariappan, Bhaskar Kosuru Bojji Raju and Daniel Ford
Eng. Proc. 2024, 73(1), 6; https://doi.org/10.3390/engproc2024073006 - 17 Oct 2024
Viewed by 96
Abstract
The COVID-19 epidemic has raised awareness of exactly how crucial it is to continuously observe issues and diagnose respiratory problems early. Although the respiratory system is the primary objective of the disease’s acute phase, subsequent complications of SARS-CoV-2 infection might trigger enduring respiratory [...] Read more.
The COVID-19 epidemic has raised awareness of exactly how crucial it is to continuously observe issues and diagnose respiratory problems early. Although the respiratory system is the primary objective of the disease’s acute phase, subsequent complications of SARS-CoV-2 infection might trigger enduring respiratory problems and symptoms, according to new research. These signs and symptoms, which collectively inflict considerable strain on healthcare systems and people’s quality of life, comprise, but are not restricted to, congestion, shortage of breath, tightness in the chest, and a decrease in lung function. Wearable technology offers a promising remedy to this persistent issue by offering continuous respiratory parameter monitoring, facilitating the early control and intervention of post-COVID-19 issues with respiration. In an effort to enhance patient outcomes and reduce expenses related to healthcare, this paper examines the possibility of using wearable technology to provide remote surveillance and the early diagnosis of respiratory problems in individuals suffering from COVID-19. In this work, a fog computing-based cost-effective smart health monitoring device is proposed for infectious disease applications. Further, the proposed device consists of three different biosensor modules, namely a MAX90614 infrared temperature sensor, a MAX30100 pulse oximeter, and a microphone sensor. All these sensor modules are connected to a fog computing device, namely a Raspberry PI microcontroller. Also, three different sensor modules were integrated with the Raspberry PI microcontroller and individuals’ physiological parameters, such as oxygen saturation (SPO2), heartbeat rate, and cough sounds, were recorded by the computing device. Additionally, a convolutional neural network (CNN)-based deep learning algorithm was coded inside the Raspberry PI and was trained with normal and COVID-19 cough sounds from the KAGGLE database. This work appears to be of high clinical significance since the developed fog computing-based smart health monitoring device is capable of identifying the presence of infectious disease through individual physiological parameters. Full article
(This article belongs to the Proceedings of The 4th International Electronic Conference on Biosensors)
Show Figures

Figure 1

19 pages, 7421 KiB  
Article
Utilizing Convolutional Neural Networks for the Effective Classification of Rice Leaf Diseases Through a Deep Learning Approach
by Salma Akter, Rashadul Islam Sumon, Haider Ali and Hee-Cheol Kim
Electronics 2024, 13(20), 4095; https://doi.org/10.3390/electronics13204095 - 17 Oct 2024
Viewed by 380
Abstract
Rice is the primary staple food in many Asian countries, and ensuring the quality of rice crops is vital for food security. Effective crop management depends on the early and precise detection of common rice diseases such as bacterial blight, blast, brown spot, [...] Read more.
Rice is the primary staple food in many Asian countries, and ensuring the quality of rice crops is vital for food security. Effective crop management depends on the early and precise detection of common rice diseases such as bacterial blight, blast, brown spot, and tungro. This work presents a convolutional neural network model for classifying rice leaf disease. Four distinct diseases, bacterial blight, blast, brown spot, and tungro, are the main targets of the model. Previously, leaf pathologies in crops were mostly identified manually using specialized equipment, which was time-consuming and inefficient. This study offers a remedy for accurately diagnosing and classifying rice leaf diseases through deep learning techniques. Using this dataset, the proposed CNN model was trained to identify complex patterns and attributes linked to each disease using its deep learning capabilities. This CNN model achieved an exceptional accuracy of 99.99%, surpassing the benchmarks set by existing state-of-the-art models. The proposed model can be a useful diagnostic and early warning system for rice leaf diseases. It could help farmers and other agricultural professionals reduce crop losses and enhance the quality of their yields. Full article
Show Figures

Figure 1

21 pages, 3741 KiB  
Article
An Efficient CNN-Based Intrusion Detection System for IoT: Use Case Towards Cybersecurity
by Amogh Deshmukh and Kiran Ravulakollu
Technologies 2024, 12(10), 203; https://doi.org/10.3390/technologies12100203 - 17 Oct 2024
Viewed by 461
Abstract
Today’s environment demands that cybersecurity be given top priority because of the increase in cyberattacks and the development of quantum computing capabilities. Traditional security measures have relied on cryptographic techniques to safeguard information systems and networks. However, with the adaptation of artificial intelligence [...] Read more.
Today’s environment demands that cybersecurity be given top priority because of the increase in cyberattacks and the development of quantum computing capabilities. Traditional security measures have relied on cryptographic techniques to safeguard information systems and networks. However, with the adaptation of artificial intelligence (AI), there is an opportunity to enhance cybersecurity through learning-based methods. IoT environments, in particular, work with lightweight systems that cannot handle the large data communications typically required by traditional intrusion detection systems (IDSs) to find anomalous patterns, making it a challenging problem. A deep learning-based framework is proposed in this study with various optimizations for automatically detecting and classifying cyberattacks. These optimizations involve dimensionality reduction, hyperparameter tuning, and feature engineering. Additionally, the framework utilizes an enhanced Convolutional Neural Network (CNN) variant called Intelligent Intrusion Detection Network (IIDNet) to detect and classify attacks efficiently. Layer optimization at the architectural level is used to improve detection performance in IIDNet using a Learning-Based Intelligent Intrusion Detection (LBIID) algorithm. The experimental study conducted in this paper uses a benchmark dataset known as UNSW-NB15 and demonstrated that IIDNet achieves an outstanding accuracy of 95.47% while significantly reducing training time and excellent scalability, outperforming many existing intrusion detection models. Full article
(This article belongs to the Special Issue IoT-Enabling Technologies and Applications)
Show Figures

Figure 1

24 pages, 12250 KiB  
Article
DMCCT: Dual-Branch Multi-Granularity Convolutional Cross-Substitution Transformer for Hyperspectral Image Classification
by Laiying Fu, Xiaoyong Chen, Yanan Xu and Xiao Li
Appl. Sci. 2024, 14(20), 9499; https://doi.org/10.3390/app14209499 - 17 Oct 2024
Viewed by 333
Abstract
In the field of hyperspectral image classification, deep learning technology, especially convolutional neural networks, has achieved remarkable progress. However, convolutional neural network models encounter challenges in hyperspectral image classification due to limitations in their receptive fields. Conversely, the global modeling capability of Transformers [...] Read more.
In the field of hyperspectral image classification, deep learning technology, especially convolutional neural networks, has achieved remarkable progress. However, convolutional neural network models encounter challenges in hyperspectral image classification due to limitations in their receptive fields. Conversely, the global modeling capability of Transformers has garnered attention in hyperspectral image classification. Nevertheless, the high computational cost and inadequate local feature extraction hinder its widespread application. In this study, we propose a novel fusion model of convolutional neural networks and Transformers to enhance performance in hyperspectral image classification, namely the dual-branch multi-granularity convolutional cross-substitution Transformer (DMCCT). The proposed model adopts a dual-branch structure to separately extract spatial and spectral features, thereby mitigating mutual interference and information loss between spectral and spatial data during feature extraction. Moreover, a multi-granularity embedding module is introduced to facilitate multi-scale and multi-level local feature extraction for spatial and spectral information. In particular, the improved convolutional cross-substitution Transformer module effectively integrates convolution and Transformer, reducing the complexity of attention operations and enhancing the accuracy of hyperspectral image classification tasks. Subsequently, the proposed method is evaluated against existing approaches using three classical datasets, namely Pavia University, Kennedy Space Center, and Indian Pines. Experimental results demonstrate the efficacy of the proposed method, achieving significant classification results on these datasets with overall classification accuracies of 98.57%, 97.96%, and 96.59%, respectively. These results establish the superiority of the proposed method in the context of hyperspectral image classification under similar experimental conditions. Full article
Show Figures

Figure 1

15 pages, 666 KiB  
Article
Comparative Analysis of AI Models for Atypical Pigmented Facial Lesion Diagnosis
by Alessandra Cartocci, Alessio Luschi, Linda Tognetti, Elisa Cinotti, Francesca Farnetani, Aimilios Lallas, John Paoli, Caterina Longo, Elvira Moscarella, Danica Tiodorovic, Ignazio Stanganelli, Mariano Suppa, Emi Dika, Iris Zalaudek, Maria Antonietta Pizzichetta, Jean Luc Perrot, Gabriele Cevenini, Ernesto Iadanza, Giovanni Rubegni, Harald Kittler, Philipp Tschandl and Pietro Rubegniadd Show full author list remove Hide full author list
Bioengineering 2024, 11(10), 1036; https://doi.org/10.3390/bioengineering11101036 - 17 Oct 2024
Viewed by 330
Abstract
Diagnosing atypical pigmented facial lesions (aPFLs) is a challenging topic for dermatologists. Accurate diagnosis of these lesions is crucial for effective patient management, especially in dermatology, where visual assessment plays a central role. Incorrect diagnoses can result in mismanagement, delays in appropriate interventions, [...] Read more.
Diagnosing atypical pigmented facial lesions (aPFLs) is a challenging topic for dermatologists. Accurate diagnosis of these lesions is crucial for effective patient management, especially in dermatology, where visual assessment plays a central role. Incorrect diagnoses can result in mismanagement, delays in appropriate interventions, and potential harm. AI, however, holds the potential to enhance diagnostic accuracy and provide reliable support to clinicians. This work aimed to evaluate and compare the effectiveness of machine learning (logistic regression of lesion features and patient metadata) and deep learning (CNN analysis of images) models in dermoscopy diagnosis and the management of aPFLs. This study involved the analysis of 1197 dermoscopic images of facial lesions excised due to suspicious and histologically confirmed malignancy, classified into seven classes (lentigo maligna—LM; lentigo maligna melanoma—LMM; atypical nevi—AN; pigmented actinic keratosis—PAK; solar lentigo—SL; seborrheic keratosis—SK; and seborrheic lichenoid keratosis—SLK). Image samples were collected through the Integrated Dermoscopy Score (iDScore) project. The statistical analysis of the dataset shows that the patients mean age was 65.5 ± 14.2, and the gender was equally distributed (580 males—48.5%; 617 females—51.5%). A total of 41.7% of the sample constituted malignant lesions (LM and LMM). Meanwhile, the benign lesions were mainly PAK (19.3%), followed by SL (22.2%), AN (10.4%), SK (4.0%), and SLK (2.3%). The lesions were mainly localised in the cheek and nose areas. A stratified analysis of the assessment provided by the enrolled dermatologists was also performed, resulting in 2445 evaluations of the 1197 images (2.1 evaluations per image on average). The physicians demonstrated higher accuracy in differentiating between malignant and benign lesions (71.2%) than in distinguishing between the seven specific diagnoses across all the images (42.9%). The logistic regression model obtained a precision of 39.1%, a sensitivity of 100%, a specificity of 33.9%, and an accuracy of 53.6% on the test set, while the CNN model showed lower sensitivity (58.2%) and higher precision (47.0%), specificity (90.8%), and accuracy (59.5%) for melanoma diagnosis. This research demonstrates how AI can enhance the diagnostic accuracy in complex dermatological cases like aPFLs by integrating AI models with clinical data and evaluating different diagnostic approaches, paving the way for more precise and scalable AI applications in dermatology, showing their critical role in improving patient management and the outcomes in dermatology. Full article
(This article belongs to the Special Issue Computer Vision and Machine Learning in Medical Applications)
Show Figures

Figure 1

Back to TopTop