Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (164)

Search Parameters:
Keywords = Gabor features

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 7535 KiB  
Article
A New Computer-Aided Diagnosis System for Breast Cancer Detection from Thermograms Using Metaheuristic Algorithms and Explainable AI
by Hanane Dihmani, Abdelmajid Bousselham and Omar Bouattane
Algorithms 2024, 17(10), 462; https://doi.org/10.3390/a17100462 - 18 Oct 2024
Viewed by 739
Abstract
Advances in the early detection of breast cancer and treatment improvements have significantly increased survival rates. Traditional screening methods, including mammography, MRI, ultrasound, and biopsies, while effective, often come with high costs and risks. Recently, thermal imaging has gained attention due to its [...] Read more.
Advances in the early detection of breast cancer and treatment improvements have significantly increased survival rates. Traditional screening methods, including mammography, MRI, ultrasound, and biopsies, while effective, often come with high costs and risks. Recently, thermal imaging has gained attention due to its minimal risks compared to mammography, although it is not widely adopted as a primary detection tool since it depends on identifying skin temperature changes and lesions. The advent of machine learning (ML) and deep learning (DL) has enhanced the effectiveness of breast cancer detection and diagnosis using this technology. In this study, a novel interpretable computer aided diagnosis (CAD) system for breast cancer detection is proposed, leveraging Explainable Artificial Intelligence (XAI) throughout its various phases. To achieve these goals, we proposed a new multi-objective optimization approach named the Hybrid Particle Swarm Optimization algorithm (HPSO) and Hybrid Spider Monkey Optimization algorithm (HSMO). These algorithms simultaneously combined the continuous and binary representations of PSO and SMO to effectively manage trade-offs between accuracy, feature selection, and hyperparameter tuning. We evaluated several CAD models and investigated the impact of handcrafted methods such as Local Binary Patterns (LBP), Histogram of Oriented Gradients (HOG), Gabor Filters, and Edge Detection. We further shed light on the effect of feature selection and optimization on feature attribution and model decision-making processes using the SHapley Additive exPlanations (SHAP) framework, with a particular emphasis on cancer classification using the DMR-IR dataset. The results of our experiments demonstrate in all trials that the performance of the model is improved. With HSMO, our models achieved an accuracy of 98.27% and F1-score of 98.15% while selecting only 25.78% of the HOG features. This approach not only boosts the performance of CAD models but also ensures comprehensive interpretability. This method emerges as a promising and transparent tool for early breast cancer diagnosis. Full article
Show Figures

Figure 1

18 pages, 4549 KiB  
Article
A Transfer Learning-Based Framework for Classifying Lymph Node Metastasis in Prostate Cancer Patients
by Suryadipto Sarkar, Teresa Wu, Matthew Harwood and Alvin C. Silva
Biomedicines 2024, 12(10), 2345; https://doi.org/10.3390/biomedicines12102345 - 15 Oct 2024
Viewed by 785
Abstract
Background: Prostate cancer is the second most common new cancer diagnosis in the United States. It is usually slow-growing, and when it is low-grade and confined to the prostate gland, it can be treated either conservatively (through active surveillance) or with surgery. However, [...] Read more.
Background: Prostate cancer is the second most common new cancer diagnosis in the United States. It is usually slow-growing, and when it is low-grade and confined to the prostate gland, it can be treated either conservatively (through active surveillance) or with surgery. However, if the cancer has spread beyond the prostate, such as to the lymph nodes, then that indicates a more aggressive cancer, and surgery may not be adequate. Methods: The challenge is that it is often difficult for radiologists reading prostate-specific imaging such as magnetic resonance images (MRIs) to differentiate malignant lymph nodes from non-malignant ones. An emerging field is the development of artificial intelligence (AI) models, including machine learning and deep learning, for medical imaging to assist in diagnostic tasks. Earlier research focused on implementing texture algorithms to extract imaging features used in classification models. More recently, researchers began studying the use of deep learning for both stand-alone feature extraction and end-to-end classification tasks. In order to tackle the challenges inherent in small datasets, this study was designed as a scalable hybrid framework utilizing pre-trained ResNet-18, a deep learning model, to extract features that were subsequently fed into a machine learning classifier to automatically identify malignant lymph nodes in patients with prostate cancer. For comparison, two texture algorithms were implemented, namely the gray-level co-occurrence matrix (GLCM) and Gabor. Results: Using an institutional prostate lymph node dataset (42 positives, 84 negatives), the proposed framework achieved an accuracy of 76.19%, a sensitivity of 79.76%, and a specificity of 69.05%. Using GLCM features, the classification achieved an accuracy of 61.90%, a sensitivity of 74.07%, and a specificity of 42.86%. Using Gabor features, the classification achieved an accuracy of 65.08%, a sensitivity of 73.47%, and a specificity of 52.50%. Conclusions: Our results demonstrate that a hybrid approach, i.e., using a pre-trainined deep learning model for feature extraction, followed by a machine learning classifier, is a viable solution. This hybrid approach is especially useful in medical-imaging-based applications with small datasets. Full article
(This article belongs to the Special Issue Advanced Cancer Diagnosis and Treatment: Second Edition)
Show Figures

Figure 1

10 pages, 3009 KiB  
Article
Unsupervised Learning for the Automatic Counting of Grains in Nanocrystals and Image Segmentation at the Atomic Resolution
by Woonbae Sohn, Taekyung Kim, Cheon Woo Moon, Dongbin Shin, Yeji Park, Haneul Jin and Hionsuck Baik
Nanomaterials 2024, 14(20), 1614; https://doi.org/10.3390/nano14201614 - 10 Oct 2024
Viewed by 694
Abstract
Identifying the grain distribution and grain boundaries of nanoparticles is important for predicting their properties. Experimental methods for identifying the crystallographic distribution, such as precession electron diffraction, are limited by their probe size. In this study, we developed an unsupervised learning method by [...] Read more.
Identifying the grain distribution and grain boundaries of nanoparticles is important for predicting their properties. Experimental methods for identifying the crystallographic distribution, such as precession electron diffraction, are limited by their probe size. In this study, we developed an unsupervised learning method by applying a Gabor filter to HAADF-STEM images at the atomic level for image segmentation and automatic counting of grains in polycrystalline nanoparticles. The methodology comprises a Gabor filter for feature extraction, non-negative matrix factorization for dimension reduction, and K-means clustering. We set the threshold distance and angle between the clusters required for the number of clusters to converge so as to automatically determine the optimal number of grains. This approach can shed new light on the nature of polycrystalline nanoparticles and their structure–property relationships. Full article
(This article belongs to the Special Issue Exploring Nanomaterials through Electron Microscopy and Spectroscopy)
Show Figures

Figure 1

23 pages, 9520 KiB  
Article
Visual Feature-Guided Diamond Convolutional Network for Finger Vein Recognition
by Qiong Yao, Dan Song, Xiang Xu and Kun Zou
Sensors 2024, 24(18), 6097; https://doi.org/10.3390/s24186097 - 20 Sep 2024
Viewed by 476
Abstract
Finger vein (FV) biometrics have garnered considerable attention due to their inherent non-contact nature and high security, exhibiting tremendous potential in identity authentication and beyond. Nevertheless, challenges pertaining to the scarcity of training data and inconsistent image quality continue to impede the effectiveness [...] Read more.
Finger vein (FV) biometrics have garnered considerable attention due to their inherent non-contact nature and high security, exhibiting tremendous potential in identity authentication and beyond. Nevertheless, challenges pertaining to the scarcity of training data and inconsistent image quality continue to impede the effectiveness of finger vein recognition (FVR) systems. To tackle these challenges, we introduce the visual feature-guided diamond convolutional network (dubbed ‘VF-DCN’), a uniquely configured multi-scale and multi-orientation convolutional neural network. The VF-DCN showcases three pivotal innovations: Firstly, it meticulously tunes the convolutional kernels through multi-scale Log-Gabor filters. Secondly, it implements a distinctive diamond-shaped convolutional kernel architecture inspired by human visual perception. This design intelligently allocates more orientational filters to medium scales, which inherently carry richer information. In contrast, at extreme scales, the use of orientational filters is minimized to simulate the natural blurring of objects at extreme focal lengths. Thirdly, the network boasts a deliberate three-layer configuration and fully unsupervised training process, prioritizing simplicity and optimal performance. Extensive experiments are conducted on four FV databases, including MMCBNU_6000, FV_USM, HKPU, and ZSC_FV. The experimental results reveal that VF-DCN achieves remarkable improvement with equal error rates (EERs) of 0.17%, 0.19%, 2.11%, and 0.65%, respectively, and Accuracy Rates (ACC) of 100%, 99.97%, 98.92%, and 99.36%, respectively. These results indicate that, compared with some existing FVR approaches, the proposed VF-DCN not only achieves notable recognition accuracy but also shows fewer number of parameters and lower model complexity. Moreover, VF-DCN exhibits superior robustness across diverse FV databases. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

25 pages, 13590 KiB  
Article
Fast and Nondestructive Proximate Analysis of Coal from Hyperspectral Images with Machine Learning and Combined Spectra-Texture Features
by Jihua Mao, Hengqian Zhao, Yu Xie, Mengmeng Wang, Pan Wang, Yaning Shi and Yusen Zhao
Appl. Sci. 2024, 14(17), 7920; https://doi.org/10.3390/app14177920 - 5 Sep 2024
Viewed by 664
Abstract
Proximate analysis, including ash, volatile matter, moisture, fixed carbon, and calorific value, is a fundamental aspect of fuel testing and serves as the primary method for evaluating coal quality, which is critical for the processing and utilization of coal. The traditional analytical methods [...] Read more.
Proximate analysis, including ash, volatile matter, moisture, fixed carbon, and calorific value, is a fundamental aspect of fuel testing and serves as the primary method for evaluating coal quality, which is critical for the processing and utilization of coal. The traditional analytical methods involve time-consuming and costly combustion processes, particularly when applied to large volumes of coal that need to be sampled in massive batches. Hyperspectral imaging is promising for the rapid and nondestructive determination of coal quality indices. In this study, a fast and nondestructive coal proximate analysis method with combined spectral-spatial features was developed using a hyperspectral imaging system in the 450–2500 nm range. The processed spectra were evaluated using PLSR, with the most effective MSC spectra selected. To reduce the spectral redundancy and improve the accuracy, the SPA, Boruta, iVISSA, and CARS algorithms were adopted to extract the characteristic wavelengths, and 16 prediction models were constructed and optimized based on the PLSR, RF, BPNN, and LSSVR algorithms within the Optuna framework for each quality indicator. For spatial information, the histogram statistics, gray-level covariance matrix, and Gabor filters were employed to extract the texture features within the characteristic wavelengths. The texture feature-based and combined spectral-texture feature-based prediction models were constructed by applying the spectral modeling strategy, respectively. Compared with the models based on spectral or texture features only, the LSSVR models with combined spectral-texture features achieved the highest prediction accuracy in all quality metrics, with Rp2 values of 0.993, 0.989, 0.979, 0.948, and 0.994 for Ash, VM, MC, FC, and CV, respectively. This study provides a technical reference for hyperspectral imaging technology as a new method for the rapid, nondestructive proximate analysis and quality assessment of coal. Full article
(This article belongs to the Section Optics and Lasers)
Show Figures

Figure 1

25 pages, 14077 KiB  
Article
Estimating Leaf Area Index in Apple Orchard by UAV Multispectral Images with Spectral and Texture Information
by Junru Yu, Yu Zhang, Zhenghua Song, Danyao Jiang, Yiming Guo, Yanfu Liu and Qingrui Chang
Remote Sens. 2024, 16(17), 3237; https://doi.org/10.3390/rs16173237 - 31 Aug 2024
Viewed by 1175
Abstract
The Leaf Area Index (LAI) strongly influences vegetation evapotranspiration and photosynthesis rates. Timely and accurately estimating the LAI is crucial for monitoring vegetation growth. The unmanned aerial vehicle (UAV) multispectral digital camera platform has been proven to be an effective tool for this [...] Read more.
The Leaf Area Index (LAI) strongly influences vegetation evapotranspiration and photosynthesis rates. Timely and accurately estimating the LAI is crucial for monitoring vegetation growth. The unmanned aerial vehicle (UAV) multispectral digital camera platform has been proven to be an effective tool for this purpose. Currently, most remote sensing estimations of LAIs focus on cereal crops, with limited research on economic crops such as apples. In this study, a method for estimating the LAI of an apple orchard by extracting spectral and texture information from UAV multispectral images was proposed. Specifically, field measurements were conducted to collect LAI data for 108 sample points during the final flowering (FF), fruit setting (FS), and fruit expansion (FE) stages of apple growth in 2023. Concurrently, UAV multispectral images were obtained to extract spectral and texture information (Gabor transform). The Support Vector Regression Recursive Feature Elimination (SVR-REF) was employed to select optimal features as inputs for constructing models to estimate the LAI. Finally, the optimal model was used for LAI mapping. The results indicate that integrating spectral and texture information effectively enhances the accuracy of LAI estimation, with the relative prediction deviation (RPD) for all models being greater than 2. The Categorical Boosting (CatBoost) model established for FF exhibits the highest accuracy, with a validation set R2, root mean square error (RMSE), and RPD of 0.867, 0.203, and 2.482, respectively. UAV multispectral imagery proves to be valuable in estimating apple orchard LAIs, offering real-time monitoring of apple growth and providing a scientific basis for orchard management. Full article
(This article belongs to the Special Issue Application of Satellite and UAV Data in Precision Agriculture)
Show Figures

Figure 1

33 pages, 30114 KiB  
Article
Exploring the Influence of Object, Subject, and Context on Aesthetic Evaluation through Computational Aesthetics and Neuroaesthetics
by Fangfu Lin, Wanni Xu, Yan Li and Wu Song
Appl. Sci. 2024, 14(16), 7384; https://doi.org/10.3390/app14167384 - 21 Aug 2024
Viewed by 929
Abstract
Background: In recent years, computational aesthetics and neuroaesthetics have provided novel insights into understanding beauty. Building upon the findings of traditional aesthetics, this study aims to combine these two research methods to explore an interdisciplinary approach to studying aesthetics. Method: Abstract artworks were [...] Read more.
Background: In recent years, computational aesthetics and neuroaesthetics have provided novel insights into understanding beauty. Building upon the findings of traditional aesthetics, this study aims to combine these two research methods to explore an interdisciplinary approach to studying aesthetics. Method: Abstract artworks were used as experimental materials. Based on traditional aesthetics and in combination, features of composition, tone, and texture were selected. Computational aesthetic methods were then employed to correspond these features to physical quantities: blank space, gray histogram, Gray Level Co-occurrence Matrix (GLCM), Local Binary Pattern (LBP), and Gabor filters. An electroencephalogram (EEG) experiment was carried out, in which participants conducted aesthetic evaluations of the experimental materials in different contexts (genuine, fake), and their EEG data were recorded to analyze the impact of various feature classes in the aesthetic evaluation process. Finally, a Support Vector Machines (SVMs) was utilized to model the feature data, Event-Related Potentials (ERPs), context data, and subjective aesthetic evaluation data. Result: Behavioral data revealed higher aesthetic ratings in the genuine context. ERP data indicated that genuine contexts elicited more negative deflections in the prefrontal lobes between 200 and 1000 ms. Class II compositions demonstrated more positive deflections in the parietal lobes at 50–120 ms, while Class I tones evoked more positive amplitudes in the occipital lobes at 200–300 ms. Gabor features showed significant variations in the parieto-occipital area at an early stage. Class II LBP elicited a prefrontal negative wave with a larger amplitude. The results of the SVM models indicated that the model incorporating aesthetic subject and context data (ACC = 0.76866) outperforms the model using only parameters of the aesthetic object (ACC = 0.68657). Conclusion: A positive context tends to provide participants with a more positive aesthetic experience, but abstract artworks may not respond to this positivity. During aesthetic evaluation, the ERP data activated by different features show a trend from global to local. The SVM model based on multimodal data fusion effectively predicts aesthetics, further demonstrating the feasibility of the combined research approach of computational aesthetics and neuroaesthetics. Full article
Show Figures

Figure 1

26 pages, 3348 KiB  
Article
Hybrid Feature Mammogram Analysis: Detecting and Localizing Microcalcifications Combining Gabor, Prewitt, GLCM Features, and Top Hat Filtering Enhanced with CNN Architecture
by Miguel Alejandro Hernández-Vázquez, Yazmín Mariela Hernández-Rodríguez, Fausto David Cortes-Rojas, Rafael Bayareh-Mancilla and Oscar Eduardo Cigarroa-Mayorga
Diagnostics 2024, 14(15), 1691; https://doi.org/10.3390/diagnostics14151691 - 5 Aug 2024
Cited by 1 | Viewed by 1262
Abstract
Breast cancer is a prevalent malignancy characterized by the uncontrolled growth of glandular epithelial cells, which can metastasize through the blood and lymphatic systems. Microcalcifications, small calcium deposits within breast tissue, are critical markers for early detection of breast cancer, especially in non-palpable [...] Read more.
Breast cancer is a prevalent malignancy characterized by the uncontrolled growth of glandular epithelial cells, which can metastasize through the blood and lymphatic systems. Microcalcifications, small calcium deposits within breast tissue, are critical markers for early detection of breast cancer, especially in non-palpable carcinomas. These microcalcifications, appearing as small white spots on mammograms, are challenging to identify due to potential confusion with other tissues. This study hypothesizes that a hybrid feature extraction approach combined with Convolutional Neural Networks (CNNs) can significantly enhance the detection and localization of microcalcifications in mammograms. The proposed algorithm employs Gabor, Prewitt, and Gray Level Co-occurrence Matrix (GLCM) kernels for feature extraction. These features are input to a CNN architecture designed with maxpooling layers, Rectified Linear Unit (ReLU) activation functions, and a sigmoid response for binary classification. Additionally, the Top Hat filter is used for precise localization of microcalcifications. The preprocessing stage includes enhancing contrast using the Volume of Interest Look-Up Table (VOI LUT) technique and segmenting regions of interest. The CNN architecture comprises three convolutional layers, three ReLU layers, and three maxpooling layers. The training was conducted using a balanced dataset of digital mammograms, with the Adam optimizer and binary cross-entropy loss function. Our method achieved an accuracy of 89.56%, a sensitivity of 82.14%, and a specificity of 91.47%, outperforming related works, which typically report accuracies around 85–87% and sensitivities between 76 and 81%. These results underscore the potential of combining traditional feature extraction techniques with deep learning models to improve the detection and localization of microcalcifications. This system may serve as an auxiliary tool for radiologists, enhancing early detection capabilities and potentially reducing diagnostic errors in mass screening programs. Full article
(This article belongs to the Special Issue Quantitative and Intelligent Analysis of Medical Imaging, 2nd Edition)
Show Figures

Figure 1

17 pages, 14796 KiB  
Article
Application of Gabor, Log-Gabor, and Adaptive Gabor Filters in Determining the Cut-Off Wavelength Shift of TFBG Sensors
by Sławomir Cięszczyk
Appl. Sci. 2024, 14(15), 6394; https://doi.org/10.3390/app14156394 - 23 Jul 2024
Viewed by 626
Abstract
Tilted fibre Bragg gratings are optical fibre structures used as sensors of various physical quantities. Their unique measurement capabilities result from the high complexity of the optical spectrum consisting of several dozen cladding mode resonances. TFBG spectra demodulation methods generate signal features that [...] Read more.
Tilted fibre Bragg gratings are optical fibre structures used as sensors of various physical quantities. Their unique measurement capabilities result from the high complexity of the optical spectrum consisting of several dozen cladding mode resonances. TFBG spectra demodulation methods generate signal features that highlight changes in the spectrum due to changes in the interacting quantities. Such methods should enable the distinction between two slightly different values of the measured quantity. The paper presents an effective method of processing the TFBG spectrum for use in measuring the refractive index of liquids. The use of Gabor and log-Gabor filters and their adaptive version eliminates the problem of discontinuity in determining the SRI value related to the existence of the cladding mode comb. The Gabor filters used make visible the shifting and fading of spectral features related to the decrease in the intensity of leaking modes. Subsequent modifications of the proposed algorithm led to an increase in the quality factor of the processed spectrum. Full article
(This article belongs to the Section Optics and Lasers)
Show Figures

Figure 1

25 pages, 2248 KiB  
Article
SCMs: Systematic Conglomerated Models for Audio Cough Signal Classification
by Sunil Kumar Prabhakar and Dong-Ok Won
Algorithms 2024, 17(7), 302; https://doi.org/10.3390/a17070302 - 8 Jul 2024
Viewed by 910
Abstract
A common and natural physiological response of the human body is cough, which tries to push air and other wastage thoroughly from the airways. Due to environmental factors, allergic responses, pollution or some diseases, cough occurs. A cough can be either dry or [...] Read more.
A common and natural physiological response of the human body is cough, which tries to push air and other wastage thoroughly from the airways. Due to environmental factors, allergic responses, pollution or some diseases, cough occurs. A cough can be either dry or wet depending on the amount of mucus produced. A characteristic feature of the cough is the sound, which is a quacking sound mostly. Human cough sounds can be monitored continuously, and so, cough sound classification has attracted a lot of interest in the research community in the last decade. In this research, three systematic conglomerated models (SCMs) are proposed for audio cough signal classification. The first conglomerated technique utilizes the concept of robust models like the Cross-Correlation Function (CCF) and Partial Cross-Correlation Function (PCCF) model, Least Absolute Shrinkage and Selection Operator (LASSO) model, elastic net regularization model with Gabor dictionary analysis and efficient ensemble machine learning techniques, the second technique utilizes the concept of stacked conditional autoencoders (SAEs) and the third technique utilizes the concept of using some efficient feature extraction schemes like Tunable Q Wavelet Transform (TQWT), sparse TQWT, Maximal Information Coefficient (MIC), Distance Correlation Coefficient (DCC) and some feature selection techniques like the Binary Tunicate Swarm Algorithm (BTSA), aggregation functions (AFs), factor analysis (FA), explanatory factor analysis (EFA) classified with machine learning classifiers, kernel extreme learning machine (KELM), arc-cosine ELM, Rat Swarm Optimization (RSO)-based KELM, etc. The techniques are utilized on publicly available datasets, and the results show that the highest classification accuracy of 98.99% was obtained when sparse TQWT with AF was implemented with an arc-cosine ELM classifier. Full article
(This article belongs to the Special Issue Quantum and Classical Artificial Intelligence)
Show Figures

Figure 1

16 pages, 3627 KiB  
Article
New Approach for Brain Tumor Segmentation Based on Gabor Convolution and Attention Mechanism
by Yuan Cao and Yinglei Song
Appl. Sci. 2024, 14(11), 4919; https://doi.org/10.3390/app14114919 - 6 Jun 2024
Viewed by 969
Abstract
In the treatment process of brain tumors, it is of great importance to develop a set of MRI image segmentation methods with high accuracy and low cost. In order to extract the feature information for each region of the brain tumor more effectively, [...] Read more.
In the treatment process of brain tumors, it is of great importance to develop a set of MRI image segmentation methods with high accuracy and low cost. In order to extract the feature information for each region of the brain tumor more effectively, this paper proposes a new model Ga-U-Net based on Gabor convolution and an attention mechanism. Based on 3D U-Net, Gabor convolution is added at the shallow layer of the encoder, which is able to learn the local structure and texture information of the tumor better. After that, the CBAM attention mechanism is added after the output of each layer of the encoder, which not only enhances the network’s ability to perceive the brain tumor boundary information but also reduces some redundant information by allocating the attention to the two dimensions of space and channel. Experimental results show that the model performs well for multiple tumor regions (WT, TC, ET) on the brain tumor dataset BraTS 2021, with Dice coefficients of 0.910, 0.897, and 0.856, respectively, which are improved by 0.3%, 2%, and 1.7% compared to the base network, the U-Net network, with an average Dice of 0.887 and an average Hausdorff distance of 9.12, all of which are better than a few other state-of-the-art deep models for biomedical image segmentation. Full article
Show Figures

Figure 1

22 pages, 3024 KiB  
Article
Augmenting Aquaculture Efficiency through Involutional Neural Networks and Self-Attention for Oplegnathus Punctatus Feeding Intensity Classification from Log Mel Spectrograms
by Usama Iqbal, Daoliang Li, Zhuangzhuang Du, Muhammad Akhter, Zohaib Mushtaq, Muhammad Farrukh Qureshi and Hafiz Abbad Ur Rehman
Animals 2024, 14(11), 1690; https://doi.org/10.3390/ani14111690 - 5 Jun 2024
Cited by 2 | Viewed by 885
Abstract
Understanding the feeding dynamics of aquatic animals is crucial for aquaculture optimization and ecosystem management. This paper proposes a novel framework for analyzing fish feeding behavior based on a fusion of spectrogram-extracted features and deep learning architecture. Raw audio waveforms are first transformed [...] Read more.
Understanding the feeding dynamics of aquatic animals is crucial for aquaculture optimization and ecosystem management. This paper proposes a novel framework for analyzing fish feeding behavior based on a fusion of spectrogram-extracted features and deep learning architecture. Raw audio waveforms are first transformed into Log Mel Spectrograms, and a fusion of features such as the Discrete Wavelet Transform, the Gabor filter, the Local Binary Pattern, and the Laplacian High Pass Filter, followed by a well-adapted deep model, is proposed to capture crucial spectral and spectral information that can help distinguish between the various forms of fish feeding behavior. The Involutional Neural Network (INN)-based deep learning model is used for classification, achieving an accuracy of up to 97% across various temporal segments. The proposed methodology is shown to be effective in accurately classifying the feeding intensities of Oplegnathus punctatus, enabling insights pertinent to aquaculture enhancement and ecosystem management. Future work may include additional feature extraction modalities and multi-modal data integration to further our understanding and contribute towards the sustainable management of marine resources. Full article
(This article belongs to the Special Issue Animal Health and Welfare in Aquaculture)
Show Figures

Figure 1

24 pages, 7093 KiB  
Article
Synthetic Aperture Radar Image Change Detection Based on Principal Component Analysis and Two-Level Clustering
by Liangliang Li, Hongbing Ma, Xueyu Zhang, Xiaobin Zhao, Ming Lv and Zhenhong Jia
Remote Sens. 2024, 16(11), 1861; https://doi.org/10.3390/rs16111861 - 23 May 2024
Cited by 10 | Viewed by 965
Abstract
Synthetic aperture radar (SAR) change detection provides a powerful tool for continuous, reliable, and objective observation of the Earth, supporting a wide range of applications that require regular monitoring and assessment of changes in the natural and built environment. In this paper, we [...] Read more.
Synthetic aperture radar (SAR) change detection provides a powerful tool for continuous, reliable, and objective observation of the Earth, supporting a wide range of applications that require regular monitoring and assessment of changes in the natural and built environment. In this paper, we introduce a novel SAR image change detection method based on principal component analysis and two-level clustering. First, two difference images of the log-ratio and mean-ratio operators are computed, then the principal component analysis fusion model is used to fuse the two difference images, and a new difference image is generated. To incorporate contextual information during the feature extraction phase, Gabor wavelets are used to obtain the representation of the difference image across multiple scales and orientations. The maximum magnitude across all orientations at each scale is then concatenated to form the Gabor feature vector. Following this, a cascading clustering algorithm is developed within this discriminative feature space by merging the first-level fuzzy c-means clustering with the second-level neighbor rule. Ultimately, the two-level combination of the changed and unchanged results produces the final change map. Five SAR datasets are used for the experiment, and the results show that our algorithm has significant advantages in SAR change detection. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

17 pages, 6884 KiB  
Article
Gradient Weakly Sensitive Multi-Source Sensor Image Registration Method
by Ronghua Li, Mingshuo Zhao, Haopeng Xue, Xinyu Li and Yuan Deng
Mathematics 2024, 12(8), 1186; https://doi.org/10.3390/math12081186 - 15 Apr 2024
Cited by 1 | Viewed by 754
Abstract
Aiming at the nonlinear radiometric differences between multi-source sensor images and coherent spot noise and other factors that lead to alignment difficulties, the registration method of gradient weakly sensitive multi-source sensor images is proposed, which does not need to extract the image gradient [...] Read more.
Aiming at the nonlinear radiometric differences between multi-source sensor images and coherent spot noise and other factors that lead to alignment difficulties, the registration method of gradient weakly sensitive multi-source sensor images is proposed, which does not need to extract the image gradient in the whole process and has rotational invariance. In the feature point detection stage, the maximum moment map is obtained by using the phase consistency transform to replace the gradient edge map for chunked Harris feature point detection, thus increasing the number of repeated feature points in the heterogeneous image. To have rotational invariance of the subsequent descriptors, a method to determine the main phase angle is proposed. The phase angle of the region near the feature point is counted, and the parabolic interpolation method is used to estimate the more accurate main phase angle under the determined interval. In the feature description stage, the Log-Gabor convolution sequence is used to construct the index map with the maximum phase amplitude, the heterogeneous image is converted to an isomorphic image, and the isomorphic image of the region around the feature point is rotated by using the main phase angle, which is in turn used to construct the feature vector with the feature point as the center by the quadratic interpolation method. In the feature matching stage, feature matching is performed by using the sum of squares of Euclidean distances as a similarity metric. Finally, after qualitative and quantitative experiments of six groups of five pairs of different multi-source sensor image alignment correct matching rates, root mean square errors, and the number of correctly matched points statistics, this algorithm is verified to have the advantage of robust accuracy compared with the current algorithms. Full article
(This article belongs to the Special Issue Applied Mathematical Modeling and Intelligent Algorithms)
Show Figures

Figure 1

22 pages, 15475 KiB  
Article
Background Subtraction for Dynamic Scenes Using Gabor Filter Bank and Statistical Moments
by Julio-Alejandro Romero-González, Diana-Margarita Córdova-Esparza, Juan Terven, Ana-Marcela Herrera-Navarro and Hugo Jiménez-Hernández
Algorithms 2024, 17(4), 133; https://doi.org/10.3390/a17040133 - 25 Mar 2024
Viewed by 1366
Abstract
This paper introduces a novel background subtraction method that utilizes texture-level analysis based on the Gabor filter bank and statistical moments. The method addresses the challenge of accurately detecting moving objects that exhibit similar color intensity variability or texture to the surrounding environment, [...] Read more.
This paper introduces a novel background subtraction method that utilizes texture-level analysis based on the Gabor filter bank and statistical moments. The method addresses the challenge of accurately detecting moving objects that exhibit similar color intensity variability or texture to the surrounding environment, which conventional methods struggle to handle effectively. The proposed method accurately distinguishes between foreground and background objects by capturing different frequency components using the Gabor filter bank and quantifying the texture level through statistical moments. Extensive experimental evaluations use datasets featuring varying lighting conditions, uniform and non-uniform textures, shadows, and dynamic backgrounds. The performance of the proposed method is compared against other existing methods using metrics such as sensitivity, specificity, and false positive rate. The experimental results demonstrate that the proposed method outperforms other methods in accuracy and robustness. It effectively handles scenarios with complex backgrounds, lighting changes, and objects that exhibit similar texture or color intensity as the background. Our method retains object structure while minimizing false detections and noise. This paper provides valuable insights into computer vision and object detection, offering a promising solution for accurate foreground detection in various applications such as video surveillance and motion tracking. Full article
Show Figures

Graphical abstract

Back to TopTop