Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (16,706)

Search Parameters:
Keywords = convolution neural network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 5181 KiB  
Article
Shift-Invariance Robustness of Convolutional Neural Networks in Side-Channel Analysis
by Marina Krček, Lichao Wu, Guilherme Perin and Stjepan Picek
Mathematics 2024, 12(20), 3279; https://doi.org/10.3390/math12203279 (registering DOI) - 18 Oct 2024
Abstract
Convolutional neural networks (CNNs) offer unrivaled performance in profiling side-channel analysis. This claim is corroborated by numerous results where CNNs break targets protected with masking and hiding countermeasures. One hiding countermeasure commonly investigated in related works is desynchronization (misalignment). The conclusions usually state [...] Read more.
Convolutional neural networks (CNNs) offer unrivaled performance in profiling side-channel analysis. This claim is corroborated by numerous results where CNNs break targets protected with masking and hiding countermeasures. One hiding countermeasure commonly investigated in related works is desynchronization (misalignment). The conclusions usually state that CNNs can break desynchronization as they are shift-invariant. This paper investigates that claim in more detail and reveals that the situation is more complex. While CNNs have certain shift-invariance, it is insufficient for commonly encountered scenarios in deep learning-based side-channel analysis. We investigate data augmentation to improve the shift-invariance and, in a more powerful version, ensembles of data augmentation. Our results show that the proposed techniques work very well and improve the attack significantly, even for an order of magnitude. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence to Cryptography)
Show Figures

Figure 1

18 pages, 8484 KiB  
Article
Feasibility of Emergency Flood Traffic Road Damage Assessment by Integrating Remote Sensing Images and Social Media Information
by Hong Zhu, Jian Meng, Jiaqi Yao and Nan Xu
ISPRS Int. J. Geo-Inf. 2024, 13(10), 369; https://doi.org/10.3390/ijgi13100369 (registering DOI) - 18 Oct 2024
Abstract
In the context of global climate change, the frequency of sudden natural disasters is increasing. Assessing traffic road damage post-disaster is crucial for emergency decision-making and disaster management. Traditional ground observation methods for evaluating traffic road damage are limited by the timeliness and [...] Read more.
In the context of global climate change, the frequency of sudden natural disasters is increasing. Assessing traffic road damage post-disaster is crucial for emergency decision-making and disaster management. Traditional ground observation methods for evaluating traffic road damage are limited by the timeliness and coverage of data updates. Relying solely on these methods does not adequately support rapid assessment and emergency management during extreme natural disasters. Social media, a major source of big data, can effectively address these limitations by providing more timely and comprehensive disaster information. Motivated by this, we utilized multi-source heterogeneous data to assess the damage to traffic roads under extreme conditions and established a new framework for evaluating traffic roads in cities prone to flood disasters caused by rainstorms. The approach involves several steps: First, the surface area affected by precipitation is extracted using a threshold method constrained by confidence intervals derived from microwave remote sensing images. Second, disaster information is collected from the Sina Weibo platform, where social media information is screened and cleaned. A quantification table for road traffic loss assessment was defined, and a social media disaster information classification model combining text convolutional neural networks and attention mechanisms (TextCNN-Attention disaster information classification) was proposed. Finally, traffic road information on social media is matched with basic geographic data, the classification of traffic road disaster risk levels is visualized, and the assessment of traffic road disaster levels is completed based on multi-source heterogeneous data. Using the “7.20” rainstorm event in Henan Province as an example, this research categorizes the disaster’s impact on traffic roads into five levels—particularly severe, severe, moderate, mild, and minimal—as derived from remote sensing image monitoring and social media information analysis. The evaluation framework for flood disaster traffic roads based on multi-source heterogeneous data provides important data support and methodological support for enhancing disaster management capabilities and systems. Full article
Show Figures

Figure 1

20 pages, 1638 KiB  
Article
GPT-Driven Radiology Report Generation with Fine-Tuned Llama 3
by Ștefan-Vlad Voinea, Mădălin Mămuleanu, Rossy Vlăduț Teică, Lucian Mihai Florescu, Dan Selișteanu and Ioana Andreea Gheonea
Bioengineering 2024, 11(10), 1043; https://doi.org/10.3390/bioengineering11101043 (registering DOI) - 18 Oct 2024
Abstract
The integration of deep learning into radiology has the potential to enhance diagnostic processes, yet its acceptance in clinical practice remains limited due to various challenges. This study aimed to develop and evaluate a fine-tuned large language model (LLM), based on Llama 3-8B, [...] Read more.
The integration of deep learning into radiology has the potential to enhance diagnostic processes, yet its acceptance in clinical practice remains limited due to various challenges. This study aimed to develop and evaluate a fine-tuned large language model (LLM), based on Llama 3-8B, to automate the generation of accurate and concise conclusions in magnetic resonance imaging (MRI) and computed tomography (CT) radiology reports, thereby assisting radiologists and improving reporting efficiency. A dataset comprising 15,000 radiology reports was collected from the University of Medicine and Pharmacy of Craiova’s Imaging Center, covering a diverse range of MRI and CT examinations made by four experienced radiologists. The Llama 3-8B model was fine-tuned using transfer-learning techniques, incorporating parameter quantization to 4-bit precision and low-rank adaptation (LoRA) with a rank of 16 to optimize computational efficiency on consumer-grade GPUs. The model was trained over five epochs using an NVIDIA RTX 3090 GPU, with intermediary checkpoints saved for monitoring. Performance was evaluated quantitatively using Bidirectional Encoder Representations from Transformers Score (BERTScore), Recall-Oriented Understudy for Gisting Evaluation (ROUGE), Bilingual Evaluation Understudy (BLEU), and Metric for Evaluation of Translation with Explicit Ordering (METEOR) metrics on a held-out test set. Additionally, a qualitative assessment was conducted, involving 13 independent radiologists who participated in a Turing-like test and provided ratings for the AI-generated conclusions. The fine-tuned model demonstrated strong quantitative performance, achieving a BERTScore F1 of 0.8054, a ROUGE-1 F1 of 0.4998, a ROUGE-L F1 of 0.4628, and a METEOR score of 0.4282. In the human evaluation, the artificial intelligence (AI)-generated conclusions were preferred over human-written ones in approximately 21.8% of cases, indicating that the model’s outputs were competitive with those of experienced radiologists. The average rating of the AI-generated conclusions was 3.65 out of 5, reflecting a generally favorable assessment. Notably, the model maintained its consistency across various types of reports and demonstrated the ability to generalize to unseen data. The fine-tuned Llama 3-8B model effectively generates accurate and coherent conclusions for MRI and CT radiology reports. By automating the conclusion-writing process, this approach can assist radiologists in reducing their workload and enhancing report consistency, potentially addressing some barriers to the adoption of deep learning in clinical practice. The positive evaluations from independent radiologists underscore the model’s potential utility. While the model demonstrated strong performance, limitations such as dataset bias, limited sample diversity, a lack of clinical judgment, and the need for large computational resources require further refinement and real-world validation. Future work should explore the integration of such models into clinical workflows, address ethical and legal considerations, and extend this approach to generate complete radiology reports. Full article
17 pages, 1085 KiB  
Article
Enhancing Brain Tumor Diagnosis with L-Net: A Novel Deep Learning Approach for MRI Image Segmentation and Classification
by Lehel Dénes-Fazakas, Levente Kovács, György Eigner and László Szilágyi
Biomedicines 2024, 12(10), 2388; https://doi.org/10.3390/biomedicines12102388 (registering DOI) - 18 Oct 2024
Abstract
Background: Brain tumors are highly complex, making their detection and classification a significant challenge in modern medical diagnostics. The accurate segmentation and classification of brain tumors from MRI images are crucial for effective treatment planning. This study aims to develop an advanced neural [...] Read more.
Background: Brain tumors are highly complex, making their detection and classification a significant challenge in modern medical diagnostics. The accurate segmentation and classification of brain tumors from MRI images are crucial for effective treatment planning. This study aims to develop an advanced neural network architecture that addresses these challenges. Methods: We propose L-net, a novel architecture combining U-net for tumor boundary segmentation and a convolutional neural network (CNN) for tumor classification. These two units are coupled such a way that the CNN classifies the MRI images based on the features extracted by the U-net while segmenting the tumor, instead of relying on the original input images. The model is trained on a dataset of 3064 high-resolution MRI images, encompassing gliomas, meningiomas, and pituitary tumors, ensuring robust performance across different tumor types. Results: L-net achieved a classification accuracy of up to 99.6%, surpassing existing models in both segmentation and classification tasks. The model demonstrated effectiveness even with lower image resolutions, making it suitable for diverse clinical settings. Conclusions: The proposed L-net model provides an accurate and unified approach to brain tumor segmentation and classification. Its enhanced performance contributes to more reliable and precise diagnosis, supporting early detection and treatment in clinical applications. Full article
(This article belongs to the Special Issue Diagnosis, Pathogenesis and Treatment of CNS Tumors)
16 pages, 2542 KiB  
Article
Deep Learning-Based Reconstruction of 3D Morphology of Geomaterial Particles from Single-View 2D Images
by Jiangpeng Zhao, Heping Xie, Cunbao Li and Yifei Liu
Materials 2024, 17(20), 5100; https://doi.org/10.3390/ma17205100 (registering DOI) - 18 Oct 2024
Abstract
The morphology of particles formed in different environments contains critical information. Thus, the rapid and effective reconstruction of their three-dimensional (3D) morphology is crucial. This study reconstructs the 3D morphology from two-dimensional (2D) images of particles using artificial intelligence (AI). More than 100,000 [...] Read more.
The morphology of particles formed in different environments contains critical information. Thus, the rapid and effective reconstruction of their three-dimensional (3D) morphology is crucial. This study reconstructs the 3D morphology from two-dimensional (2D) images of particles using artificial intelligence (AI). More than 100,000 particles were sampled from three sources: naturally formed particles (desert sand), manufactured particles (lunar soil simulant), and numerically generated digital particles. A deep learning approach based on a voxel representation of the morphology and multi-dimensional convolutional neural networks was proposed to rapidly upscale and reconstruct particle morphology. The trained model was tested using the three particle types and evaluated using different multi-scale morphological descriptors. The results demonstrated that the statistical properties of the morphological descriptors were consistent for the real 3D particles and those derived from the 2D images and the model. This finding confirms the model’s validity and generalizability in upscaling and reconstructing diverse particle samples. This study provides a method for generating 3D numerical representations of geological particles, facilitating in-depth analysis of properties, such as mechanical behavior and transport characteristics, from 2D images. Full article
21 pages, 20740 KiB  
Article
Pattern Recognition in the Processing of Electromyographic Signals for Selected Expressions of Polish Sign Language
by Anna Filipowska, Wojciech Filipowski, Julia Mieszczanin, Katarzyna Bryzik, Maciej Henkel, Emilia Skwarek, Paweł Raif, Szymon Sieciński, Rafał Doniec, Barbara Mika, Julia Bodak, Piotr Ferst, Marcin Pieniążek, Kamil Pilarski and Marcin Grzegorzek
Sensors 2024, 24(20), 6710; https://doi.org/10.3390/s24206710 (registering DOI) - 18 Oct 2024
Abstract
Gesture recognition has become a significant part of human–machine interaction, particularly when verbal interaction is not feasible. The rapid development of biomedical sensing and machine learning algorithms, including electromyography (EMG) and convolutional neural networks (CNNs), has enabled the interpretation of sign languages, including [...] Read more.
Gesture recognition has become a significant part of human–machine interaction, particularly when verbal interaction is not feasible. The rapid development of biomedical sensing and machine learning algorithms, including electromyography (EMG) and convolutional neural networks (CNNs), has enabled the interpretation of sign languages, including the Polish Sign Language, based on EMG signals. The objective was to classify the game control gestures and Polish Sign Language gestures recorded specifically for this study using two different data acquisition systems: BIOPAC MP36 and MyoWare 2.0. We compared the classification performance of various machine learning algorithms, with a particular emphasis on CNNs on the dataset of EMG signals representing 24 gestures, recorded using both types of EMG sensors. The results (98.324% versus ≤7.8571% and 95.5307% versus ≤10.2697% of accuracy for CNNs and other classifiers in data recorded with BIOPAC MP36 and MyoWare, respectively) indicate that CNNs demonstrate superior accuracy. These results suggest the feasibility of using lower-cost sensors for effective gesture classification and the viability of integrating affordable EMG-based technologies into broader gesture recognition frameworks, providing a cost-effective solution for real-world applications. The dataset created during the study offers a basis for future studies on EMG-based recognition of Polish Sign Language. Full article
(This article belongs to the Special Issue Wearable Sensors, Robotic Systems and Assistive Devices)
25 pages, 3905 KiB  
Article
An Efficient Ship Detection Method Based on YOLO and Ship Wakes Using High-Resolution Optical Jilin1 Satellite Imagery
by Fangli Mou, Zide Fan, Yunping Ge, Lei Wang and Xinming Li
Sensors 2024, 24(20), 6708; https://doi.org/10.3390/s24206708 (registering DOI) - 18 Oct 2024
Abstract
In this study, we propose a practical and efficient scheme for ship detection in remote sensing imagery. Our method is developed using both ship body detection and ship wake detection and combines deep learning and feature-based image processing. A deep convolutional neural network [...] Read more.
In this study, we propose a practical and efficient scheme for ship detection in remote sensing imagery. Our method is developed using both ship body detection and ship wake detection and combines deep learning and feature-based image processing. A deep convolutional neural network is used to achieve ship body detection, and a feature-based processing method is proposed to detect ship wakes. For better analysis, we model the sea region and evaluate the quality of the image. Generally, the wake detection result is used to assist ship detection and obtain the sailing direction. Conventional methods cannot detect ships that are covered by clouds or outside the image boundary. The method proposed in this paper uses the wake to detect such ships, with a certain level of confidence and low false alarm probability in detection. Practical aspects such as the method’s applicability and time efficiency are considered in our method for real applications. We demonstrate the effectiveness of our method in a real remote sensing dataset. The results show that over 93.5% of ships and over 70% of targets with no visible ship body can be successfully detected. This illustrates that the proposed detection framework can fill the gap regarding the detection of sailing ships in a remote sensing image. Full article
(This article belongs to the Section Remote Sensors)
20 pages, 10555 KiB  
Article
Cloud Detection Using a UNet3+ Model with a Hybrid Swin Transformer and EfficientNet (UNet3+STE) for Very-High-Resolution Satellite Imagery
by Jaewan Choi, Doochun Seo, Jinha Jung, Youkyung Han, Jaehong Oh and Changno Lee
Remote Sens. 2024, 16(20), 3880; https://doi.org/10.3390/rs16203880 (registering DOI) - 18 Oct 2024
Abstract
It is necessary to extract and recognize the cloud regions presented in imagery to generate satellite imagery as analysis-ready data (ARD). In this manuscript, we proposed a new deep learning model to detect cloud areas in very-high-resolution (VHR) satellite imagery by fusing two [...] Read more.
It is necessary to extract and recognize the cloud regions presented in imagery to generate satellite imagery as analysis-ready data (ARD). In this manuscript, we proposed a new deep learning model to detect cloud areas in very-high-resolution (VHR) satellite imagery by fusing two deep learning architectures. The proposed UNet3+ model with a hybrid Swin Transformer and EfficientNet (UNet3+STE) was based on the structure of UNet3+, with the encoder sequentially combining EfficientNet based on mobile inverted bottleneck convolution (MBConv) and the Swin Transformer. By sequentially utilizing convolutional neural networks (CNNs) and transformer layers, the proposed algorithm aimed to extract the local and global information of cloud regions effectively. In addition, the decoder used MBConv to restore the spatial information of the feature map extracted by the encoder and adopted the deep supervision strategy of UNet3+ to enhance the model’s performance. The proposed model was trained using the open dataset derived from KOMPSAT-3 and 3A satellite imagery and conducted a comparative evaluation with the state-of-the-art (SOTA) methods on fourteen test datasets at the product level. The experimental results confirmed that the proposed UNet3+STE model outperformed the SOTA methods and demonstrated the most stable precision, recall, and F1 score values with fewer parameters and lower complexity. Full article
(This article belongs to the Special Issue Deep Learning Techniques Applied in Remote Sensing)
Show Figures

Figure 1

34 pages, 8862 KiB  
Article
A Novel Detection Transformer Framework for Ship Detection in Synthetic Aperture Radar Imagery Using Advanced Feature Fusion and Polarimetric Techniques
by Mahmoud Ahmed, Naser El-Sheimy and Henry Leung
Remote Sens. 2024, 16(20), 3877; https://doi.org/10.3390/rs16203877 - 18 Oct 2024
Abstract
Ship detection in synthetic aperture radar (SAR) imagery faces significant challenges due to the limitations of traditional methods, such as convolutional neural network (CNN) and anchor-based matching approaches, which struggle with accurately detecting smaller targets as well as adapting to varying environmental conditions. [...] Read more.
Ship detection in synthetic aperture radar (SAR) imagery faces significant challenges due to the limitations of traditional methods, such as convolutional neural network (CNN) and anchor-based matching approaches, which struggle with accurately detecting smaller targets as well as adapting to varying environmental conditions. These methods, relying on either intensity values or single-target characteristics, often fail to enhance the signal-to-clutter ratio (SCR) and are prone to false detections due to environmental factors. To address these issues, a novel framework is introduced that leverages the detection transformer (DETR) model along with advanced feature fusion techniques to enhance ship detection. This feature enhancement DETR (FEDETR) module manages clutter and improves feature extraction through preprocessing techniques such as filtering, denoising, and applying maximum and median pooling with various kernel sizes. Furthermore, it combines metrics like the line spread function (LSF), peak signal-to-noise ratio (PSNR), and F1 score to predict optimal pooling configurations and thus enhance edge sharpness, image fidelity, and detection accuracy. Complementing this, the weighted feature fusion (WFF) module integrates polarimetric SAR (PolSAR) methods such as Pauli decomposition, coherence matrix analysis, and feature volume and helix scattering (Fvh) components decomposition, along with FEDETR attention maps, to provide detailed radar scattering insights that enhance ship response characterization. Finally, by integrating wave polarization properties, the ability to distinguish and characterize targets is augmented, thereby improving SCR and facilitating the detection of weakly scattered targets in SAR imagery. Overall, this new framework significantly boosts DETR’s performance, offering a robust solution for maritime surveillance and security. Full article
(This article belongs to the Special Issue Target Detection with Fully-Polarized Radar)
Show Figures

Figure 1

22 pages, 16009 KiB  
Article
Lightweight Multi-Domain Fusion Model for Through-Wall Human Activity Recognition Using IR-UWB Radar
by Ling Huang, Dong Lei, Bowen Zheng, Guiping Chen, Huifeng An and Mingxuan Li
Appl. Sci. 2024, 14(20), 9522; https://doi.org/10.3390/app14209522 - 18 Oct 2024
Abstract
Impulse radio ultra-wideband (IR-UWB) radar, operating in the low-frequency band, can penetrate walls and utilize its high range resolution to recognize different human activities. Complex deep neural networks have demonstrated significant performance advantages in classifying radar spectrograms of various actions, but at the [...] Read more.
Impulse radio ultra-wideband (IR-UWB) radar, operating in the low-frequency band, can penetrate walls and utilize its high range resolution to recognize different human activities. Complex deep neural networks have demonstrated significant performance advantages in classifying radar spectrograms of various actions, but at the cost of a substantial computational overhead. In response, this paper proposes a lightweight model named TG2-CAFNet. First, clutter suppression and time–frequency analysis are used to obtain range–time and micro-Doppler feature maps of human activities. Then, leveraging GhostV2 convolution, a lightweight feature extraction module, TG2, suitable for radar spectrograms is constructed. Using a parallel structure, the features of the two spectrograms are extracted separately. Finally, to further explore the correlation between the two spectrograms and enhance the feature representation capabilities, an improved nonlinear fusion method called coordinate attention fusion (CAF) is proposed based on attention feature fusion (AFF). This method extends the adaptive weighting fusion of AFF to a spatial distribution, effectively capturing the subtle spatial relationships between the two radar spectrograms. Experiments showed that the proposed method achieved a high degree of model lightweightness, while also achieving a recognition accuracy of 99.1%. Full article
Show Figures

Figure 1

30 pages, 39939 KiB  
Article
Urban Color Perception and Sentiment Analysis Based on Deep Learning and Street View Big Data
by Mingyang Yu, Xiangyu Zheng, Pinrui Qin, Weikang Cui and Qingrui Ji
Appl. Sci. 2024, 14(20), 9521; https://doi.org/10.3390/app14209521 - 18 Oct 2024
Abstract
The acceleration of urbanization has resulted in a heightened awareness of the impacts of urban environments on residents’ emotional states. This present study focuses on the Lixia District of Jinan City. By using urban street view big data and deep learning methods, we [...] Read more.
The acceleration of urbanization has resulted in a heightened awareness of the impacts of urban environments on residents’ emotional states. This present study focuses on the Lixia District of Jinan City. By using urban street view big data and deep learning methods, we undertook a detailed analysis of the impacts of urban color features on residents’ emotional perceptions. In particular, a substantial corpus of street scene image data was extracted and processed. This was performed using a deep convolutional neural network (DCNN) and semantic segmentation technology (PSPNet), which enabled the simulation and prediction of the subjective perception of the urban environment by humans. Furthermore, the color complexity and coordination in the street scene were quantified and combined with residents’ emotional feedback to carry out a multi-dimensional analysis. The findings revealed that color complexity and coordination were significant elements influencing residents’ emotional perceptions. A high color complexity is visually appealing, but can lead to fatigue, discomfort, and boredom; a moderate complexity stimulates vitality and pleasure; high levels of regional harmony and aesthetics can increase perceptions of beauty and security; and low levels of coordination can increase feelings of depression. The environmental characteristics of different areas and differences in the daily activities of residents resulted in regional differences regarding the impacts of color features on emotional perception. This study corroborates the assertion that environmental color coordination has the capacity to enhance residents’ emotions, thereby providing an important reference point for urban planning. Planning should be based on the functional characteristics of the region, and color complexity and coordination should be reasonably regulated to optimize the emotional experiences of residents. Differentiated color management enhances urban aesthetics, livability, and residents’ happiness and promotes sustainable development. In the future, the influences of color and environmental factors on emotions can be explored in depth, with a view to assist in the formulation of fine urban design. Full article
Show Figures

Figure 1

13 pages, 1012 KiB  
Article
Lightweight Optic Disc and Optic Cup Segmentation Based on MobileNetv3 Convolutional Neural Network
by Yuanqiong Chen, Zhijie Liu, Yujia Meng and Jianfeng Li
Biomimetics 2024, 9(10), 637; https://doi.org/10.3390/biomimetics9100637 - 18 Oct 2024
Abstract
Glaucoma represents a significant global contributor to blindness. Accurately segmenting the optic disc (OD) and optic cup (OC) to obtain precise CDR is essential for effective screening. However, existing convolutional neural network (CNN)-based segmentation techniques are often limited by high computational demands and [...] Read more.
Glaucoma represents a significant global contributor to blindness. Accurately segmenting the optic disc (OD) and optic cup (OC) to obtain precise CDR is essential for effective screening. However, existing convolutional neural network (CNN)-based segmentation techniques are often limited by high computational demands and long inference times. This paper proposes an efficient end-to-end method for OD and OC segmentation, utilizing the lightweight MobileNetv3 network as the core feature-extraction module. Our approach combines boundary branches with adversarial learning, to achieve multi-label segmentation of the OD and OC. We validated our proposed approach across three public available datasets: Drishti-GS, RIM-ONE-r3, and REFUGE. The outcomes reveal that the Dice coefficients for the segmentation of OD and OC within these datasets are 0.974/0.900, 0.966/0.875, and 0.962/0.880, respectively. Additionally, our method substantially lowers computational complexity and inference time, thereby enabling efficient and precise segmentation of the optic disc and optic cup. Full article
Show Figures

Figure 1

14 pages, 3525 KiB  
Article
Deep Learning-Based Gender Recognition in Cherry Valley Ducks Through Sound Analysis
by Guofeng Han, Yujing Liu, Jiawen Cai, Enze Duan, Zefeng Shi, Shida Zhao, Lianfei Huo, Huixin Wang and Zongchun Bai
Animals 2024, 14(20), 3017; https://doi.org/10.3390/ani14203017 - 18 Oct 2024
Abstract
Gender recognition is an important part of the duck industry. Currently, the gender identification of ducks mainly relies on manual labor, which is highly labor-intensive. This study aims to propose a novel method for distinguishing between males and females based on the characteristic [...] Read more.
Gender recognition is an important part of the duck industry. Currently, the gender identification of ducks mainly relies on manual labor, which is highly labor-intensive. This study aims to propose a novel method for distinguishing between males and females based on the characteristic sound parameters for day-old ducks. The effective data from the sounds of day-old ducks were recorded and extracted using the endpoint detection method. The 12-dimensional Mel-frequency cepstral coefficients (MFCCs) with first-order and second-order difference coefficients in the effective sound signals of the ducks were calculated, and a total of 36-dimensional feature vectors were obtained. These data were used as input information to train three classification models, include a backpropagation neural network (BPNN), a deep neural network (DNN), and a convolutional neural network (CNN). The training results show that the accuracies of the BPNN, DNN, and CNN were 83.87%, 83.94%, and 84.15%, respectively, and that the three classification models could identify the sounds of male and female ducks. The prediction results showed that the prediction accuracies of the BPNN, DNN, and CNN were 93.33%, 91.67%, and 95.0%, respectively, which shows that the scheme for distinguishing between male and female ducks via sound had high accuracy. Moreover, the CNN demonstrated the best recognition effect. The method proposed in this study can provide some support for developing an efficient technique for gender identification in duck production. Full article
Show Figures

Figure 1

29 pages, 14557 KiB  
Article
Compressive Strength Prediction of Fly Ash-Based Concrete Using Single and Hybrid Machine Learning Models
by Haiyu Li, Heungjin Chung, Zhenting Li and Weiping Li
Buildings 2024, 14(10), 3299; https://doi.org/10.3390/buildings14103299 - 18 Oct 2024
Abstract
The compressive strength of concrete is a crucial parameter in structural design, yet its determination in a laboratory setting is both time-consuming and expensive. The prediction of compressive strength in fly ash-based concrete can be accelerated through the use of machine learning algorithms [...] Read more.
The compressive strength of concrete is a crucial parameter in structural design, yet its determination in a laboratory setting is both time-consuming and expensive. The prediction of compressive strength in fly ash-based concrete can be accelerated through the use of machine learning algorithms with artificial intelligence, which can effectively address the problems associated with this process. This paper presents the most innovative model algorithms established based on artificial intelligence technology. These include three single models—a fully connected neural network model (FCNN), a convolutional neural network model (CNN), and a transformer model (TF)—and three hybrid models—FCNN + CNN, TF + FCNN, and TF + CNN. A total of 471 datasets were employed in the experiments, comprising 7 input features: cement (C), fly ash (FA), water (W), superplasticizer (SP), coarse aggregate (CA), fine aggregate (S), and age (D). Six models were subsequently applied to predict the compressive strength (CS) of fly ash-based concrete. Furthermore, the loss function curves, assessment indexes, linear correlation coefficient, and the related literature indexes of each model were employed for comparison. This analysis revealed that the FCNN + CNN model exhibited the highest prediction accuracy, with the following metrics: R2 = 0.95, MSE = 14.18, MAE = 2.32, SMAPE = 0.1, and R = 0.973. Additionally, SHAP was utilized to elucidate the significance of the model parameter features. The findings revealed that C and D exerted the most substantial influence on the model prediction outcomes, followed by W and FA. Nevertheless, CA, S, and SP demonstrated comparatively minimal influence. Finally, a GUI interface for predicting compressive strength was developed based on six models and nonlinear functional relationships, and a criterion for minimum strength was derived by comparison and used to optimize a reasonable mixing ratio, thus achieving a fast data-driven interaction that was concise and reliable. Full article
(This article belongs to the Section Building Materials, and Repair & Renovation)
Show Figures

Figure 1

32 pages, 4351 KiB  
Article
Enhancing Brain Tumor Detection Through Custom Convolutional Neural Networks and Interpretability-Driven Analysis
by Kavinda Ashan Kulasinghe Wasalamuni Dewage, Raza Hasan, Bacha Rehman and Salman Mahmood
Information 2024, 15(10), 653; https://doi.org/10.3390/info15100653 - 18 Oct 2024
Abstract
Brain tumor detection is crucial for effective treatment planning and improved patient outcomes. However, existing methods often face challenges, such as limited interpretability and class imbalance in medical-imaging data. This study presents a novel, custom Convolutional Neural Network (CNN) architecture, specifically designed to [...] Read more.
Brain tumor detection is crucial for effective treatment planning and improved patient outcomes. However, existing methods often face challenges, such as limited interpretability and class imbalance in medical-imaging data. This study presents a novel, custom Convolutional Neural Network (CNN) architecture, specifically designed to address these issues by incorporating interpretability techniques and strategies to mitigate class imbalance. We trained and evaluated four CNN models (proposed CNN, ResNetV2, DenseNet201, and VGG16) using a brain tumor MRI dataset, with oversampling techniques and class weighting employed during training. Our proposed CNN achieved an accuracy of 94.51%, outperforming other models in regard to precision, recall, and F1-Score. Furthermore, interpretability was enhanced through gradient-based attribution methods and saliency maps, providing valuable insights into the model’s decision-making process and fostering collaboration between AI systems and clinicians. This approach contributes a highly accurate and interpretable framework for brain tumor detection, with the potential to significantly enhance diagnostic accuracy and personalized treatment planning in neuro-oncology. Full article
(This article belongs to the Special Issue Artificial Intelligence and Data Science for Health)
Show Figures

Figure 1

Back to TopTop