Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,199)

Search Parameters:
Keywords = similar pixels

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 2289 KiB  
Article
Automatic Watershed Segmentation of Cancerous Lesions in Unsupervised Breast Histology Images
by Vincent Majanga and Ernest Mnkandla
Appl. Sci. 2024, 14(22), 10394; https://doi.org/10.3390/app142210394 - 12 Nov 2024
Viewed by 91
Abstract
Segmentation of nuclei in histology images is key in analyzing and quantifying morphology changes of nuclei features and tissue structures. Conventional diagnosis, segmenting, and detection methods have relied heavily on the manual-visual inspection of histology images. These methods are only effective on clearly [...] Read more.
Segmentation of nuclei in histology images is key in analyzing and quantifying morphology changes of nuclei features and tissue structures. Conventional diagnosis, segmenting, and detection methods have relied heavily on the manual-visual inspection of histology images. These methods are only effective on clearly visible cancerous lesions on histology images thus limited in their performance due to the complexity of tissue structures in histology images. Hence, early detection of breast cancer is key for treatment and profits from Computer-Aided-Diagnostic (CAD) systems introduced to efficiently and automatically segment and detect nuclei cells in pathology. This paper proposes, an automatic watershed segmentation method of cancerous lesions in unsupervised human breast histology images. Firstly, this approach pre-processes data through various augmentation methods to increase the size of dataset images, then a stain normalization technique is applied to these augmented images to isolate nuclei features from tissue structures. Secondly, data enhancement techniques namely; erosion, dilation, and distance transform are used to highlight foreground and background pixels while removing unwanted regions from the highlighted nuclei objects on the image. Consequently, the connected components method groups these highlighted pixel components with similar intensity values and, assigns them to their relevant labeled component binary mask. Once all binary masked groups have been determined, a deep-learning recurrent neural network from the Keras architecture uses this information to automatically segment nuclei objects with cancerous lesions and their edges on the image via watershed filling. This segmentation method is evaluated on an unsupervised, augmented human breast cancer histology dataset of 11,151 images. This proposed method produced a significant evaluation result of 98% F1-accuracy score. Full article
Show Figures

Figure 1

20 pages, 17265 KiB  
Article
Satellite-Observed Hydrothermal Conditions Control the Effects of Soil and Atmospheric Drought on Peak Vegetation Growth on the Tibetan Plateau
by Zhengliang Qiu, Longxiang Tang, Xiaoyue Wang, Yunfei Zhang, Jianbo Tan, Jun Yue and Shaobo Xia
Remote Sens. 2024, 16(22), 4163; https://doi.org/10.3390/rs16224163 - 8 Nov 2024
Viewed by 309
Abstract
Recent research has demonstrated that global warming significantly enhances peak vegetation growth on the Tibetan Plateau (TP), underscoring the influence of climatic factors on vegetation dynamics. Nevertheless, the effects of different drought types on peak vegetation growth remain underexplored. This study utilized satellite-derived [...] Read more.
Recent research has demonstrated that global warming significantly enhances peak vegetation growth on the Tibetan Plateau (TP), underscoring the influence of climatic factors on vegetation dynamics. Nevertheless, the effects of different drought types on peak vegetation growth remain underexplored. This study utilized satellite-derived gross primary productivity (GPP) and the normalized difference vegetation index (NDVI) to assess the impacts of soil moisture (SM) and vapor pressure deficit (VPD) on peak vegetation growth (GPPmax and NDVImax) across the TP from 2001 to 2022. Our findings indicate that NDVImax and GPPmax exhibited increasing trends in most regions, displaying similar spatial patterns, with 65.28% of pixels showing an increase in NDVImax and 72.98% in GPPmax. In contrast, the trend for SM primarily showed a decrease (80.86%), while VPD showed an increasing trend (74.75%). Through partial correlation analysis and ridge regression, we found that peak vegetation growth was significantly affected by SM or VPD in nearly 20% of the study areas, although the magnitude of these effects varied considerably. Furthermore, we revealed that hydrothermal conditions modulated the responses of peak vegetation growth to SM and VPD. In regions with annual precipitation less than 650 mm and an annual mean temperature below 10 °C, decreased SM and increased VPD generally inhibited peak vegetation growth. Conversely, in warm and humid areas, lower SM and higher VPD promoted peak vegetation growth. These findings are crucial for deepening our understanding of vegetation phenology and its future responses to climate change. Full article
Show Figures

Figure 1

25 pages, 7487 KiB  
Article
Design of Shape Forming Elements for Architected Composites via Bayesian Optimization and Genetic Algorithms: A Concept Evaluation
by David O. Kazmer, Rebecca H. Olanrewaju, David C. Elbert and Thao D. Nguyen
Materials 2024, 17(21), 5339; https://doi.org/10.3390/ma17215339 - 31 Oct 2024
Viewed by 295
Abstract
This article presents the first use of shape forming elements (SFEs) to produce architected composites from multiple materials in an extrusion process. Each SFE contains a matrix of flow channels connecting input and output ports, where materials are routed between corresponding ports. The [...] Read more.
This article presents the first use of shape forming elements (SFEs) to produce architected composites from multiple materials in an extrusion process. Each SFE contains a matrix of flow channels connecting input and output ports, where materials are routed between corresponding ports. The mathematical operations of rotation and shifting are described, and design automation is explored using Bayesian optimization and genetic algorithms to select fifty or more parameters for minimizing two objective functions. The first objective aims to match a target cross-section by minimizing the pixel-by-pixel error, which is weighted with the structural similarity index (SSIM). The second objective seeks to maximize information content by minimizing the SSIM relative to a white image. Satisfactory designs are achieved with better objective function values observed in rectangular rather than square flow channels. Validation extrusion of modeling clay demonstrates that while SFEs impose complex material transformations, they do not achieve the material distributions predicted by the digital model. Using the SSIM for results comparison, initial stages yielded SSIM values near 0.8 between design and simulation, indicating a good initial match. However, the control of material processing tended to decline with successive SFE processing with the SSIM of the extruded output dropping to 0.023 relative to the design intent. Flow simulations more closely replicated the observed structures with SSIM values around 0.4 but also failed to predict the intended cross-sections. The evaluation highlights the need for advanced modeling techniques to enhance the predictive accuracy and functionality of SFEs for biomedical, energy storage, and structural applications. Full article
(This article belongs to the Special Issue Manufacturing, Characterization and Modeling of Advanced Materials)
Show Figures

Graphical abstract

20 pages, 5537 KiB  
Article
TTMGNet: Tree Topology Mamba-Guided Network Collaborative Hierarchical Incremental Aggregation for Change Detection
by Hongzhu Wang, Zhaoyi Ye, Chuan Xu, Liye Mei, Cheng Lei and Du Wang
Remote Sens. 2024, 16(21), 4068; https://doi.org/10.3390/rs16214068 - 31 Oct 2024
Viewed by 366
Abstract
Change detection (CD) identifies surface changes by analyzing bi-temporal remote sensing (RS) images of the same region and is essential for effective urban planning, ensuring the optimal allocation of resources, and supporting disaster management efforts. However, deep-learning-based CD methods struggle with background noise [...] Read more.
Change detection (CD) identifies surface changes by analyzing bi-temporal remote sensing (RS) images of the same region and is essential for effective urban planning, ensuring the optimal allocation of resources, and supporting disaster management efforts. However, deep-learning-based CD methods struggle with background noise and pseudo-changes due to local receptive field limitations or computing resource constraints, which limits long-range dependency capture and feature integration, normally resulting in fragmented detections and high false positive rates. To address these challenges, we propose a tree topology Mamba-guided network (TTMGNet) based on Mamba architecture, which combines the Mamba architecture for effectively capturing global features, a unique tree topology structure for retaining fine local details, and a hierarchical feature fusion mechanism that enhances multi-scale feature integration and robustness against noise. Specifically, the a Tree Topology Mamba Feature Extractor (TTMFE) leverages the similarity of pixels to generate minimum spanning tree (MST) topology sequences, guiding information aggregation and transmission. This approach utilizes a Tree Topology State Space Model (TTSSM) to embed spatial and positional information while preserving the global feature extraction capability, thereby retaining local features. Subsequently, the Hierarchical Incremental Aggregation Module is utilized to gradually align and merge features from deep to shallow layers to facilitate hierarchical feature integration. Through residual connections and cross-channel attention (CCA), HIAM enhances the interaction between neighboring feature maps, ensuring that critical features are retained and effectively utilized during the fusion process, thereby enabling more accurate detection results in CD. The proposed TTMGNet achieved F1 scores of 92.31% on LEVIR-CD, 90.94% on WHU-CD, and 77.25% on CL-CD, outperforming current mainstream methods in suppressing the impact of background noise and pseudo-change and more accurately identifying change regions. Full article
Show Figures

Figure 1

16 pages, 3470 KiB  
Article
YOLOv8-Based Estimation of Estrus in Sows Through Reproductive Organ Swelling Analysis Using a Single Camera
by Iyad Almadani, Mohammed Abuhussein and Aaron L. Robinson
Digital 2024, 4(4), 898-913; https://doi.org/10.3390/digital4040044 - 27 Oct 2024
Viewed by 538
Abstract
Accurate and efficient estrus detection in sows is crucial in modern agricultural practices to ensure optimal reproductive health and successful breeding outcomes. A non-contact method using computer vision to detect a change in a sow’s vulva size holds great promise for automating and [...] Read more.
Accurate and efficient estrus detection in sows is crucial in modern agricultural practices to ensure optimal reproductive health and successful breeding outcomes. A non-contact method using computer vision to detect a change in a sow’s vulva size holds great promise for automating and enhancing this critical process. However, achieving precise and reliable results depends heavily on maintaining a consistent camera distance during image capture. Variations in camera distance can lead to erroneous estrus estimations, potentially resulting in missed breeding opportunities or false positives. To address this challenge, we propose a robust six-step methodology, accompanied by three stages of evaluation. First, we carefully annotated masks around the vulva to ensure an accurate pixel perimeter calculation of its shape. Next, we meticulously identified keypoints on the sow’s vulva, which enabled precise tracking and analysis of its features. We then harnessed the power of machine learning to train our model using annotated images, which facilitated keypoint detection and segmentation with the state-of-the-art YOLOv8 algorithm. By identifying the keypoints, we performed precise calculations of the Euclidean distances: first, between each labium (horizontal distance), and second, between the clitoris and the perineum (vertical distance). Additionally, by segmenting the vulva’s size, we gained valuable insights into its shape, which helped with performing precise perimeter measurements. Equally important was our effort to calibrate the camera using monocular depth estimation. This calibration helped establish a functional relationship between the measurements on the image (such as the distances between the labia and from the clitoris to the perineum, and the vulva perimeter) and the depth distance to the camera, which enabled accurate adjustments and calibration for our analysis. Lastly, we present a classification method for distinguishing between estrus and non-estrus states in subjects based on the pixel width, pixel length, and perimeter measurements. The method calculated the Euclidean distances between a new data point and reference points from two datasets: “estrus data” and “not estrus data”. Using custom distance functions, we computed the distances for each measurement dimension and aggregated them to determine the overall similarity. The classification process involved identifying the three nearest neighbors of the datasets and employing a majority voting mechanism to assign a label. A new data point was classified as “estrus” if the majority of the nearest neighbors were labeled as estrus; otherwise, it was classified as “non-estrus”. This method provided a robust approach for automated classification, which aided in more accurate and efficient detection of the estrus states. To validate our approach, we propose three evaluation stages. In the first stage, we calculated the Mean Squared Error (MSE) between the ground truth keypoints of the labia distance and the distance between the predicted keypoints, and we performed the same calculation for the distance between the clitoris and perineum. Then, we provided a quantitative analysis and performance comparison, including a comparison between our previous U-Net model and our new YOLOv8 segmentation model. This comparison focused on each model’s performance in terms of accuracy and speed, which highlighted the advantages of our new approach. Lastly, we evaluated the estrus–not-estrus classification model by defining the confusion matrix. By using this comprehensive approach, we significantly enhanced the accuracy of estrus detection in sows while effectively mitigating human errors and resource wastage. The automation and optimization of this critical process hold the potential to revolutionize estrus detection in agriculture, which will contribute to improved reproductive health management and elevate breeding outcomes to new heights. Through extensive evaluation and experimentation, our research aimed to demonstrate the transformative capabilities of computer vision techniques, paving the way for more advanced and efficient practices in the agricultural domain. Full article
Show Figures

Figure 1

18 pages, 3655 KiB  
Article
Investigating the Role of Cover-Crop Spectra for Vineyard Monitoring from Airborne and Spaceborne Remote Sensing
by Michael Williams, Niall G. Burnside, Matthew Brolly and Chris B. Joyce
Remote Sens. 2024, 16(21), 3942; https://doi.org/10.3390/rs16213942 - 23 Oct 2024
Viewed by 526
Abstract
The monitoring of grape quality parameters within viticulture using airborne remote sensing is an increasingly important aspect of precision viticulture. Airborne remote sensing allows high volumes of spatial consistent data to be collected with improved efficiency over ground-based surveys. Spectral data can be [...] Read more.
The monitoring of grape quality parameters within viticulture using airborne remote sensing is an increasingly important aspect of precision viticulture. Airborne remote sensing allows high volumes of spatial consistent data to be collected with improved efficiency over ground-based surveys. Spectral data can be used to understand the characteristics of vineyards, including the characteristics and health of the vines. Within viticultural remote sensing, the use of cover-crop spectra for monitoring is often overlooked due to the perceived noise it generates within imagery. However, within viticulture, the cover crop is a widely used and important management tool. This study uses multispectral data acquired by a high-resolution uncrewed aerial vehicle (UAV) and Sentinel-2 MSI to explore the benefit that cover-crop pixels could have for grape yield and quality monitoring. This study was undertaken across three growing seasons in the southeast of England, at a large commercial wine producer. The site was split into a number of vineyards, with sub-blocks for different vine varieties and rootstocks. Pre-harvest multispectral UAV imagery was collected across three vineyard parcels. UAV imagery was radiometrically corrected and stitched to create orthomosaics (red, green, and near-infrared) for each vineyard and survey date. Orthomosaics were segmented into pure cover-cropuav and pure vineuav pixels, removing the impact that mixed pixels could have upon analysis, with three vegetation indices (VIs) constructed from the segmented imagery. Sentinel-2 Level 2a bottom of atmosphere scenes were also acquired as close to UAV surveys as possible. In parallel, the yield and quality surveys were undertaken one to two weeks prior to harvest. Laboratory refractometry was performed to determine the grape total acid, total soluble solids, alpha amino acids, and berry weight. Extreme gradient boosting (XGBoost v2.1.1) was used to determine the ability of remote sensing data to predict the grape yield and quality parameters. Results suggested that pure cover-cropuav was a successful predictor of grape yield and quality parameters (range of R2 = 0.37–0.45), with model evaluation results comparable to pure vineuav and Sentinel-2 models. The analysis also showed that, whilst the structural similarity between the both UAV and Sentinel-2 data was high, the cover crop is the most influential spectral component within the Sentinel-2 data. This research presents novel evidence for the ability of cover-cropuav to predict grape yield and quality. Moreover, this finding then provides a mechanism which explains the success of the Sentinel-2 modelling of grape yield and quality. For growers and wine producers, creating grape yield and quality prediction models through moderate-resolution satellite imagery would be a significant innovation. Proving more cost-effective than UAV monitoring for large vineyards, such methodologies could also act to bring substantial cost savings to vineyard management. Full article
Show Figures

Figure 1

20 pages, 6388 KiB  
Article
Extraction of Winter Wheat Planting Plots with Complex Structures from Multispectral Remote Sensing Images Based on the Modified Segformer Model
by Chunshan Wang, Shuo Yang, Penglei Zhu and Lijie Zhang
Agronomy 2024, 14(10), 2433; https://doi.org/10.3390/agronomy14102433 - 20 Oct 2024
Viewed by 538
Abstract
As one of the major global food crops, the monitoring and management of the winter wheat planting area is of great significance for agricultural production and food security worldwide. Today, the development of high-resolution remote sensing imaging technology has provided rich sources of [...] Read more.
As one of the major global food crops, the monitoring and management of the winter wheat planting area is of great significance for agricultural production and food security worldwide. Today, the development of high-resolution remote sensing imaging technology has provided rich sources of data for extracting the visual planting information of winter wheat. However, the existing research mostly focuses on extracting the planting plots that have a simple terrain structure. In the face of diverse terrain features combining mountainous areas, plains, and saline alkali land, as well as small-scale but complex planting structures, the extraction of planting plots through remote sensing imaging is subjected to great challenges in terms of recognition accuracy and model complexity. In this paper, we propose a modified Segformer model for extracting winter wheat planting plots with complex structures in rural areas based on the 0.8 m high-resolution multispectral data obtained from the Gaofen-2 satellite, which significantly improves the extraction accuracy and efficiency under complex conditions. In the encoder and decoder of this method, new modules were developed for the purpose of optimizing the feature extraction and fusion process. Specifically, the improvement measures of the proposed method include: (1) The MixFFN module in the original Segformer model is replaced with the Multi-Scale Feature Fusion Fully-connected Network (MSF-FFN) module, which enhances the model’s representation ability in handling complex terrain features through multi-scale feature extraction and position embedding convolution; furthermore, the DropPath mechanism is introduced to reduce the possibility of overfitting while improving the model’s generalization ability. (2) In the decoder part, after fusing features at four different scales, a CoordAttention module is added, which can precisely locate important regions with enhanced features in the images by utilizing the coordinate attention mechanism, therefore further improving the model’s extraction accuracy. (3) The model’s input data are strengthened by incorporating multispectral indices, which are also conducive to the improvement of the overall extraction accuracy. The experimental results show that the accuracy rate of the modified Segformer model in extracting winter wheat planting plots is significantly increased compared to traditional segmentation models, with the mean Intersection over Union (mIOU) and mean Pixel Accuracy (mPA) reaching 89.88% and 94.67%, respectively (an increase of 1.93 and 1.23 percentage points, respectively, compared to the baseline model). Meanwhile, the parameter count and computational complexity are significantly reduced compared to other similar models. Furthermore, when multispectral indices are input into the model, the mIOU and mPA reach 90.97% and 95.16%, respectively (an increase of 3.02 and 1.72 percentage points, respectively, compared to the baseline model). Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

25 pages, 13404 KiB  
Article
Drone SAR Imaging for Monitoring an Active Landslide Adjacent to the M25 at Flint Hall Farm
by Anthony Carpenter, James A. Lawrence, Philippa J. Mason, Richard Ghail and Stewart Agar
Remote Sens. 2024, 16(20), 3874; https://doi.org/10.3390/rs16203874 - 18 Oct 2024
Viewed by 811
Abstract
Flint Hall Farm in Godstone, Surrey, UK, is situated adjacent to the London Orbital Motorway, or M25, and contains several landslide systems which pose a significant geohazard risk to this critical infrastructure. The site has been routinely monitored by geotechnical engineers following a [...] Read more.
Flint Hall Farm in Godstone, Surrey, UK, is situated adjacent to the London Orbital Motorway, or M25, and contains several landslide systems which pose a significant geohazard risk to this critical infrastructure. The site has been routinely monitored by geotechnical engineers following a landslide that encroached onto the hard shoulder in December 2000; current in situ instrumentation includes inclinometers and piezoelectric sensors. Interferometric Synthetic Aperture Radar (InSAR) is an active remote sensing technique that can quantify millimetric rates of Earth surface and structural deformation, typically utilising satellite data, and is ideal for monitoring landslide movements. We have developed the hardware and software for an Unmanned Aerial Vehicle (UAV), or drone radar system, for improved operational flexibility and spatial–temporal resolutions in the InSAR data. The hardware payload includes an industrial-grade DJI drone, a high-performance Ettus Software Defined Radar (SDR), and custom Copper Clad Laminate (CCL) radar horn antennas. The software utilises Frequency Modulated Continuous Wave (FMCW) radar at 5.4 GHz for raw data collection and a Range Migration Algorithm (RMA) for focusing the data into a Single Look Complex (SLC) Synthetic Aperture Radar (SAR) image. We present the first SAR image acquired using the drone radar system at Flint Hall Farm, which provides an improved spatial resolution compared to satellite SAR. Discrete targets on the landslide slope, such as corner reflectors and the in situ instrumentation, are visible as bright pixels, with their size and positioning as expected; the surrounding grass and vegetation appear as natural speckles. Drone SAR imaging is an emerging field of research, given the necessary and recent technological advancements in drones and SDR processing power; as such, this is a novel achievement, with few authors demonstrating similar systems. Ongoing and future work includes repeat-pass SAR data collection and developing the InSAR processing chain for drone SAR data to provide meaningful deformation outputs for the landslides and other geotechnical hazards and infrastructure. Full article
Show Figures

Figure 1

14 pages, 2710 KiB  
Article
SPDepth: Enhancing Self-Supervised Indoor Monocular Depth Estimation via Self-Propagation
by Xiaotong Guo, Huijie Zhao, Shuwei Shao, Xudong Li, Baochang Zhang and Na Li
Future Internet 2024, 16(10), 375; https://doi.org/10.3390/fi16100375 - 16 Oct 2024
Viewed by 494
Abstract
Due to the existence of low-textured areas in indoor scenes, some self-supervised depth estimation methods have specifically designed sparse photometric consistency losses and geometry-based losses. However, some of the loss terms cannot supervise all the pixels, which limits the performance of these methods. [...] Read more.
Due to the existence of low-textured areas in indoor scenes, some self-supervised depth estimation methods have specifically designed sparse photometric consistency losses and geometry-based losses. However, some of the loss terms cannot supervise all the pixels, which limits the performance of these methods. Some approaches introduce an additional optical flow network to provide dense correspondences supervision, but overload the loss function. In this paper, we propose to perform depth self-propagation based on feature self-similarities, where high-accuracy depths are propagated from supervised pixels to unsupervised ones. The enhanced self-supervised indoor monocular depth estimation network is called SPDepth. Since depth self-similarities are significant in a local range, a local window self-attention module is embedded at the end of the network to propagate depths in a window. The depth of a pixel is weighted using the feature correlation scores with other pixels in the same window. The effectiveness of self-propagation mechanism is demonstrated in the experiments on the NYU Depth V2 dataset. The root-mean-squared error of SPDepth is 0.585 and the δ1 accuracy is 77.6%. Zero-shot generalization studies are also conducted on the 7-Scenes dataset and provide a more comprehensive analysis about the application characteristics of SPDepth. Full article
(This article belongs to the Special Issue Machine Learning Techniques for Computer Vision)
Show Figures

Figure 1

30 pages, 30880 KiB  
Article
Development of a New Non-Destructive Analysis Method in Cultural Heritage with Artificial Intelligence
by Bengin Bilici Genc, Erkan Bostanci, Bekir Eskici, Hakan Erten, Berna Caglar Eryurt, Koray Acici, Didem Ketenoglu and Tunc Asuroglu
Electronics 2024, 13(20), 4039; https://doi.org/10.3390/electronics13204039 - 14 Oct 2024
Viewed by 633
Abstract
Cultural assets are all movable and immovable assets that have been the subject of social life in historical periods, have unique scientific and cultural value, and are located above ground, underground or underwater. Today, the fact that most of the analyses conducted to [...] Read more.
Cultural assets are all movable and immovable assets that have been the subject of social life in historical periods, have unique scientific and cultural value, and are located above ground, underground or underwater. Today, the fact that most of the analyses conducted to understand the technologies of these assets require sampling and that non-destructive methods that allow analysis without taking samples are costly is a problem for cultural heritage workers. In this study, which was prepared to find solutions to national and international problems, it is aimed to develop a non-destructive, cost-minimizing and easy-to-use analysis method. Since this article aimed to develop methodology, the materials were prepared for preliminary research purposes. Therefore, it was limited to four primary colors. These four primary colors were red and yellow ochre, green earth, Egyptian blue and ultramarine blue. These pigments were used with different binders. The produced paints were photographed in natural and artificial light at different light intensities and brought to a 256 × 256 pixel size, and then trained on support vector machine, convolutional neural network, densely connected convolutional network, residual network 50 and visual geometry group 19 models. It was asked whether the trained VGG19 model could classify the paints used in archaeological and artistic works analyzed with instrumental methods in the literature with their real identities. As a result of the test, the model was able to classify paints in artworks from photographs non-destructively with a 99% success rate, similar to the result of the McNemar test. Full article
Show Figures

Figure 1

30 pages, 6496 KiB  
Article
Enhancement Method Based on Multi-Strategy Improved Pelican Optimization Algorithm and Application to Low-Illumination Forest Canopy Images
by Xiaohan Zhao, Liangkuan Zhu, Jingyu Wang and Alaa M. E. Mohamed
Forests 2024, 15(10), 1783; https://doi.org/10.3390/f15101783 - 11 Oct 2024
Viewed by 725
Abstract
Enhancement is a crucial step in the field of image processing, as it significantly improves image analysis and understanding. One of the most commonly used methods for image contrast enhancement is the incomplete beta function (IBF). However, the key challenge lies in determining [...] Read more.
Enhancement is a crucial step in the field of image processing, as it significantly improves image analysis and understanding. One of the most commonly used methods for image contrast enhancement is the incomplete beta function (IBF). However, the key challenge lies in determining the optimal parameters for the IBF. This paper introduces a multi-strategy improved pelican optimization algorithm (MIPOA) to address the low-illumination color image enhancement problem. The MIPOA algorithm utilizes a nonlinear decreasing coefficient to boost the exploration ability and convergence speed, whereas the Hardy–Weinberg principle compensates for the unsound exploitation mechanism. Additionally, the diversity variation operation improves the ability of the algorithm to escape local optimal solutions. The performance of the proposed MIPOA algorithm was evaluated using a benchmark function and was found to outperform five variant algorithms in extensive comparisons. To further harness the potential of the MIPOA algorithm, the authors propose a low-light forest canopy image enhancement method based on the MIPOA algorithm. The MIPOA algorithm searches for the optimal parameters of the IBF, leading to fast contrast enhancement of the image. The segmented gamma correction function is designed to enhance the brightness of the low-light forest canopy images. In determining the optimal parameters of IBF, the MIPOA algorithm demonstrates superior performance compared to other intelligent algorithms in the feature similarity index (FSIM), entropy, and contrast improvement index (CII) of 75%, 58.33%, and 75%, respectively. The proposed MIPOA-based enhancement method achieves a moderate pixel mean and surpasses the conventional enhancement method with an average gradient of 91.67%. The experimental results indicate that the MIPOA effectively addresses the limitations of low optimization accuracy in IBF parameters, and the enhancement method based on the MIPOA provides a more efficacious approach for enhancing low-light forest canopy images. Full article
(This article belongs to the Special Issue New Development of Smart Forestry: Machine and Automation)
Show Figures

Figure 1

19 pages, 14507 KiB  
Article
High-Precision Multi-Object Tracking in Satellite Videos via Pixel-Wise Adaptive Feature Enhancement
by Gang Wan, Zhijuan Su, Yitian Wu, Ningbo Guo, Dianwei Cong, Zhanji Wei, Wei Liu and Guoping Wang
Sensors 2024, 24(19), 6489; https://doi.org/10.3390/s24196489 - 9 Oct 2024
Viewed by 670
Abstract
In this paper, we focus on the multi-target tracking (MOT) task in satellite videos. To achieve efficient and accurate tracking, we propose a transformer-distillation-based end-to-end joint detection and tracking (JDT) method. Specifically, (1) considering that targets in satellite videos usually have small scales [...] Read more.
In this paper, we focus on the multi-target tracking (MOT) task in satellite videos. To achieve efficient and accurate tracking, we propose a transformer-distillation-based end-to-end joint detection and tracking (JDT) method. Specifically, (1) considering that targets in satellite videos usually have small scales and are shot from a bird’s-eye view, we propose a pixel-wise transformer-based feature distillation module through which useful object representations are learned via pixel-wise distillation using a strong teacher detection network; (2) targets in satellite videos, such as airplanes, ships, and vehicles, usually have similar appearances, so we propose a temperature-controllable key feature learning objective function, and by highlighting the learning of similar features during distilling, the tracking accuracy for such objects can be further improved; (3) we propose a method that is based on an end-to-end network but simultaneously learns from a highly precise teacher network and tracking head during training so that the tracking accuracy of the end-to-end network can be improved via distillation without compromising efficiency. The experimental results on three recently released publicly available datasets demonstrated the superior performance of the proposed method for satellite videos. The proposed method achieved over 90% overall tracking performance on the AIR-MOT dataset. Full article
Show Figures

Figure 1

29 pages, 6970 KiB  
Article
Traumatic Brain Injury Structure Detection Using Advanced Wavelet Transformation Fusion Algorithm with Proposed CNN-ViT
by Abdullah, Ansar Siddique, Zulaikha Fatima and Kamran Shaukat
Information 2024, 15(10), 612; https://doi.org/10.3390/info15100612 - 6 Oct 2024
Viewed by 668
Abstract
Detecting Traumatic Brain Injuries (TBI) through imaging remains challenging due to limited sensitivity in current methods. This study addresses the gap by proposing a novel approach integrating deep-learning algorithms and advanced image-fusion techniques to enhance detection accuracy. The method combines contextual and visual [...] Read more.
Detecting Traumatic Brain Injuries (TBI) through imaging remains challenging due to limited sensitivity in current methods. This study addresses the gap by proposing a novel approach integrating deep-learning algorithms and advanced image-fusion techniques to enhance detection accuracy. The method combines contextual and visual models to effectively assess injury status. Using a dataset of repeat mild TBI (mTBI) cases, we compared various image-fusion algorithms: PCA (89.5%), SWT (89.69%), DCT (89.08%), HIS (83.3%), and averaging (80.99%). Our proposed hybrid model achieved a significantly higher accuracy of 98.78%, demonstrating superior performance. Metrics including Dice coefficient (98%), sensitivity (97%), and specificity (98%) verified that the strategy is efficient in improving image quality and feature extraction. Additional validations with “entropy”, “average pixel intensity”, “standard deviation”, “correlation coefficient”, and “edge similarity measure” confirmed the robustness of the fused images. The hybrid CNN-ViT model, integrating curvelet transform features, was trained and validated on a comprehensive dataset of 24 types of brain injuries. The overall accuracy was 99.8%, with precision, recall, and F1-score of 99.8%. The “average PSNR” was 39.0 dB, “SSIM” was 0.99, and MI was 1.0. Cross-validation across five folds proved the model’s “dependability” and “generalizability”. In conclusion, this study introduces a promising method for TBI detection, leveraging advanced image-fusion and deep-learning techniques, significantly enhancing medical imaging and diagnostic capabilities for brain injuries. Full article
(This article belongs to the Special Issue Real-World Applications of Machine Learning Techniques)
Show Figures

Figure 1

14 pages, 5605 KiB  
Article
3D Multi-Phase Sub-Pixel PSF Estimation Based on Space Debris Detection System
by Fan Bu, Dalei Yao and Yan Wen
Photonics 2024, 11(10), 933; https://doi.org/10.3390/photonics11100933 - 3 Oct 2024
Viewed by 529
Abstract
The distribution of diffuse spot energy can be used to sensitively evaluate the aberrations and defects of optical systems. Therefore, the objective and quantitative measurement of diffuse spot parameters is an important means to control the detection quality of space debris detection systems. [...] Read more.
The distribution of diffuse spot energy can be used to sensitively evaluate the aberrations and defects of optical systems. Therefore, the objective and quantitative measurement of diffuse spot parameters is an important means to control the detection quality of space debris detection systems. At present, the existing optical system dispersion measurement method can only judge whether the energy distribution meets the index. However, these methods ca not provide an objective quantitative basis to guide the installation process. To solve this problem, a mathematical simulation model of 3D multi-phase sub-pixel PSF distribution is proposed. According to the relation between the CCD target plane and the theoretical image plane (focal plane, defocus, and deflection), the diffuse spot distribution of the optical system is simulated with different phase combinations. Then, Pearson Correlation Coefficient (PCC) is used to evaluate the matching similarity of the diffuse spot image. The simulation results show that when the PCC is greater than 0.96, the distribution of the two diffuse spots can be identified as matching. This also confirms the accuracy of the proposed PSF model. Then, the focusing deviation of the system being tested can be analyzed according to the phase size of the diffuse spot simulation image. This method can quickly and accurately guide the focal surface installation and testing of the system. Therefore, the purpose of improving the detection accuracy of space debris is achieved. It also provides a quantitative basis for the engineering application of optical detection systems in the future. Full article
Show Figures

Figure 1

24 pages, 14371 KiB  
Article
An Enhanced Transportation System for People of Determination
by Uma Perumal, Fathe Jeribi and Mohammed Hameed Alhameed
Sensors 2024, 24(19), 6411; https://doi.org/10.3390/s24196411 - 3 Oct 2024
Viewed by 700
Abstract
Visually Impaired Persons (VIPs) have difficulty in recognizing vehicles used for navigation. Additionally, they may not be able to identify the bus to their desired destination. However, the bus bay in which the designated bus stops has not been analyzed in the existing [...] Read more.
Visually Impaired Persons (VIPs) have difficulty in recognizing vehicles used for navigation. Additionally, they may not be able to identify the bus to their desired destination. However, the bus bay in which the designated bus stops has not been analyzed in the existing literature. Thus, a guidance system for VIPs that identifies the correct bus for transportation is presented in this paper. Initially, speech data indicating the VIP’s destination are pre-processed and converted to text. Next, utilizing the Arctan Gradient-activated Recurrent Neural Network (ArcGRNN) model, the number of bays at the location is detected with the help of a Global Positioning System (GPS), input text, and bay location details. Then, the optimal bay is chosen from the detected bays by utilizing the Experienced Perturbed Bacteria Foraging Triangular Optimization Algorithm (EPBFTOA), and an image of the selected bay is captured and pre-processed. Next, the bus is identified utilizing a You Only Look Once (YOLO) series model. Utilizing the Sub-pixel Shuffling Convoluted Encoder–ArcGRNN Decoder (SSCEAD) framework, the text is detected and segmented for the buses identified in the image. From the segmented output, the text is extracted, based on the destination and route of the bus. Finally, regarding the similarity value with respect to the VIP’s destination, a decision is made utilizing the Multi-characteristic Non-linear S-Curve-Fuzzy Rule (MNC-FR). This decision informs the bus conductor about the VIP, such that the bus can be stopped appropriately to pick them up. During testing, the proposed system selected the optimal bay in 247,891 ms, which led to deciding the bus stop for the VIP with a fuzzification time of 34,197 ms. Thus, the proposed model exhibits superior performance over those utilized in prevailing works. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Back to TopTop