Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,753)

Search Parameters:
Keywords = YOLOv8

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 32403 KiB  
Article
Identification and Localization of Wind Turbine Blade Faults Using Deep Learning
by Mason Davis, Edwin Nazario Dejesus, Mohammad Shekaramiz, Joshua Zander and Majid Memari
Appl. Sci. 2024, 14(14), 6319; https://doi.org/10.3390/app14146319 (registering DOI) - 19 Jul 2024
Viewed by 88
Abstract
This study addresses the challenges inherent in the maintenance and inspection of wind turbines through the application of deep learning methodologies for fault detection on Wind Turbine Blades (WTBs). Specifically, this research focuses on defect detection on the blades of small-scale WTBs due [...] Read more.
This study addresses the challenges inherent in the maintenance and inspection of wind turbines through the application of deep learning methodologies for fault detection on Wind Turbine Blades (WTBs). Specifically, this research focuses on defect detection on the blades of small-scale WTBs due to the unavailability of commercial wind turbines. This research compared popular object localization architectures, YOLO and Mask R-CNN, to identify the most effective model to detect common WTB defects, including cracks, holes, and erosion. YOLOv9 C emerged as the most effective model, with the highest scores of mAP50 and mAP50-95 of 0.849 and 0.539, respectively. Modifications to Mask R-CNN, specifically integrating a ResNet18-FPN network, reduced computational complexity by 32 layers and achieved a mAP50 of 0.8415. The findings highlight the potential of deep learning and computer vision in improving WTB fault analysis and inspection. Full article
(This article belongs to the Section Green Sustainable Science and Technology)
27 pages, 10826 KiB  
Article
CRLNet: A Multimodal Peach Detection Network Based on Cooperative Asymptotic Enhancement and the Fusion of Granularity Refinement
by Jiahao Liu, Chaoying He, Mingfang Wang, Yichu Jiang, Manman Sun, Miying Yan and Mingfang He
Plants 2024, 13(14), 1980; https://doi.org/10.3390/plants13141980 (registering DOI) - 19 Jul 2024
Viewed by 121
Abstract
Accurate peach detection is essential for automated agronomic management, such as mechanical peach harvesting. However, ubiquitous occlusion makes identifying peaches from complex backgrounds extremely challenging. In addition, it is difficult to capture fine-grained peach features from a single RGB image, which can suffer [...] Read more.
Accurate peach detection is essential for automated agronomic management, such as mechanical peach harvesting. However, ubiquitous occlusion makes identifying peaches from complex backgrounds extremely challenging. In addition, it is difficult to capture fine-grained peach features from a single RGB image, which can suffer from light and noise in scenarios with dense small target clusters and extreme light. To solve these problems, this study proposes a multimodal detector, called CRLNet, based on RGB and depth images. First, YOLOv9 was extended to design a backbone network that can extract RGB and depth features in parallel from an image. Second, to address the problem of information fusion bias, the Rough–Fine Hybrid Attention Fusion Module (RFAM) was designed to combine the advantageous information of different modes while suppressing the hollow noise at the edge of the peach. Finally, a Transformer-based Local–Global Joint Enhancement Module (LGEM) was developed to jointly enhance the local and global features of peaches using information from different modalities in order to enhance the percentage of information about the target peaches and remove the interference of redundant background information. CRLNet was trained on the Peach dataset and evaluated against other state-of-the-art methods; the model achieved an mAP50 of 97.1%. In addition, CRLNet also achieved an mAP50 of 92.4% in generalized experiments, validating its strong generalization capability. These results provide valuable insights for peach and other outdoor fruit multimodal detection. Full article
Show Figures

Figure 1

22 pages, 16539 KiB  
Article
BY-SLAM: Dynamic Visual SLAM System Based on BEBLID and Semantic Information Extraction
by Daixian Zhu, Peixuan Liu, Qiang Qiu, Jiaxin Wei and Ruolin Gong
Sensors 2024, 24(14), 4693; https://doi.org/10.3390/s24144693 (registering DOI) - 19 Jul 2024
Viewed by 103
Abstract
SLAM is a critical technology for enabling autonomous navigation and positioning in unmanned vehicles. Traditional visual simultaneous localization and mapping algorithms are built upon the assumption of a static scene, overlooking the impact of dynamic targets within real-world environments. Interference from dynamic targets [...] Read more.
SLAM is a critical technology for enabling autonomous navigation and positioning in unmanned vehicles. Traditional visual simultaneous localization and mapping algorithms are built upon the assumption of a static scene, overlooking the impact of dynamic targets within real-world environments. Interference from dynamic targets can significantly degrade the system’s localization accuracy or even lead to tracking failure. To address these issues, we propose a dynamic visual SLAM system named BY-SLAM, which is based on BEBLID and semantic information extraction. Initially, the BEBLID descriptor is introduced to describe Oriented FAST feature points, enhancing both feature point matching accuracy and speed. Subsequently, FasterNet replaces the backbone network of YOLOv8s to expedite semantic information extraction. By using the results of DBSCAN clustering object detection, a more refined semantic mask is obtained. Finally, by leveraging the semantic mask and epipolar constraints, dynamic feature points are discerned and eliminated, allowing for the utilization of only static feature points for pose estimation and the construction of a dense 3D map that excludes dynamic targets. Experimental evaluations are conducted on both the TUM RGB-D dataset and real-world scenarios and demonstrate the effectiveness of the proposed algorithm at filtering out dynamic targets within the scenes. On average, the localization accuracy for the TUM RGB-D dataset improves by 95.53% compared to ORB-SLAM3. Comparative analyses against classical dynamic SLAM systems further corroborate the improvement in localization accuracy, map readability, and robustness achieved by BY-SLAM. Full article
(This article belongs to the Section Navigation and Positioning)
30 pages, 7576 KiB  
Article
Improved DeepSORT-Based Object Tracking in Foggy Weather for AVs Using Sematic Labels and Fused Appearance Feature Network
by Isaac Ogunrinde and Shonda Bernadin
Sensors 2024, 24(14), 4692; https://doi.org/10.3390/s24144692 (registering DOI) - 19 Jul 2024
Viewed by 119
Abstract
The presence of fog in the background can prevent small and distant objects from being detected, let alone tracked. Under safety-critical conditions, multi-object tracking models require faster tracking speed while maintaining high object-tracking accuracy. The original DeepSORT algorithm used YOLOv4 for the detection [...] Read more.
The presence of fog in the background can prevent small and distant objects from being detected, let alone tracked. Under safety-critical conditions, multi-object tracking models require faster tracking speed while maintaining high object-tracking accuracy. The original DeepSORT algorithm used YOLOv4 for the detection phase and a simple neural network for the deep appearance descriptor. Consequently, the feature map generated loses relevant details about the track being matched with a given detection in fog. Targets with a high degree of appearance similarity on the detection frame are more likely to be mismatched, resulting in identity switches or track failures in heavy fog. We propose an improved multi-object tracking model based on the DeepSORT algorithm to improve tracking accuracy and speed under foggy weather conditions. First, we employed our camera-radar fusion network (CR-YOLOnet) in the detection phase for faster and more accurate object detection. We proposed an appearance feature network to replace the basic convolutional neural network. We incorporated GhostNet to take the place of the traditional convolutional layers to generate more features and reduce computational complexities and costs. We adopted a segmentation module and fed the semantic labels of the corresponding input frame to add rich semantic information to the low-level appearance feature maps. Our proposed method outperformed YOLOv5 + DeepSORT with a 35.15% increase in multi-object tracking accuracy, a 32.65% increase in multi-object tracking precision, a speed increase by 37.56%, and identity switches decreased by 46.81%. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

20 pages, 8422 KiB  
Article
Welding Seam Tracking and Inspection Robot Based on Improved YOLOv8s-Seg Model
by Minghu Zhao, Xinru Liu, Kaihang Wang, Zishen Liu, Qi Dong, Pengfei Wang and Yaoheng Su
Sensors 2024, 24(14), 4690; https://doi.org/10.3390/s24144690 (registering DOI) - 19 Jul 2024
Viewed by 144
Abstract
A weld is the main connection form of special equipment, and a weld is also the most vulnerable part of special equipment. Therefore, an effective detection of a weld is of great significance to improve the safety of special equipment. The traditional inspection [...] Read more.
A weld is the main connection form of special equipment, and a weld is also the most vulnerable part of special equipment. Therefore, an effective detection of a weld is of great significance to improve the safety of special equipment. The traditional inspection method is not only time-consuming and labor-intensive, but also expensive. The welding seam tracking and inspection robot can greatly improve the inspection efficiency and save on inspection costs. Therefore, this paper proposes a welding seam tracking and inspection robot based on YOLOv8s-seg. Firstly, the MobileNetV3 lightweight backbone network is used to replace the backbone part of YOLOv8s-seg to reduce the model parameters. Secondly, we reconstruct C2f and prune the number of output channels of the new building module C2fGhost. Finally, in order to make up for the precision loss caused by the lightweight model, we add an EMA attention mechanism after each detection layer in the neck part of the model. The experimental results show that the accuracy of weld recognition reaches 97.8%, and the model size is only 4.88 MB. The improved model is embedded in Jetson nano, a robot control system for seam tracking and detection, and TensorRT is used to accelerate the reasoning of the model. The total reasoning time from image segmentation to path fitting is only 54 ms, which meets the real-time requirements of the robot for seam tracking and detection, and realizes the path planning of the robot for inspecting the seam efficiently and accurately. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

19 pages, 5673 KiB  
Article
A Live Detecting System for Strain Clamps of Transmission Lines Based on Dual UAVs’ Cooperation
by Zhiwei Jia, Yongkang Ouyang, Chao Feng, Shaosheng Fan, Zheng Liu and Chenhao Sun
Drones 2024, 8(7), 333; https://doi.org/10.3390/drones8070333 (registering DOI) - 19 Jul 2024
Viewed by 136
Abstract
Strain clamps are critical components in high-voltage overhead transmission lines, and detection of their defects becomes an important part of regular inspection of transmission lines. A dual UAV (unmanned aerial vehicle) system was proposed to detect strain clamps in multiple split-phase conductors. The [...] Read more.
Strain clamps are critical components in high-voltage overhead transmission lines, and detection of their defects becomes an important part of regular inspection of transmission lines. A dual UAV (unmanned aerial vehicle) system was proposed to detect strain clamps in multiple split-phase conductors. The main UAV was equipped with a digital radiography (DR) imaging device, a mechanical arm, and an edge intelligence module with visual sensors. The slave UAV was equipped with a digital imaging board and visual sensors. A workflow was proposed for this dual UAV system. Target detection and distance detection of the strain clamps, as well as detection of the defects of strain clamps in DR images, are the main procedures of this workflow. To satisfy the demands of UAV-borne and real-time deployment, the improved YOLOv8-TR algorithm was proposed for the detection of strain clamps (the mAP@50 was 60.9%), and the KD-ResRPA algorithm is used for detecting defects in DR images (the average AUCROC of the three datasets was 82.7%). Field experiments validated the suitability of our dual UAV-based system for charged detection of strain clamps in double split-phase conductors, demonstrating its potential for practical application in live detecting systems. Full article
(This article belongs to the Special Issue Embodied Artificial Intelligence Systems for UAVs)
Show Figures

Figure 1

21 pages, 2276 KiB  
Article
Optimizing Automated Detection of Cross-Recessed Screws in Laptops Using a Neural Network
by Nicholas M. DiFilippo, Musa K. Jouaneh and Alexander D. Jedson
Appl. Sci. 2024, 14(14), 6301; https://doi.org/10.3390/app14146301 (registering DOI) - 19 Jul 2024
Viewed by 144
Abstract
This paper investigates varying the operating conditions of a neural network in a robotic system using a low-cost webcam to achieve optimal settings in order to detect crossed-recess screws on laptops, a necessary step in the realization of automated disassembly systems. A study [...] Read more.
This paper investigates varying the operating conditions of a neural network in a robotic system using a low-cost webcam to achieve optimal settings in order to detect crossed-recess screws on laptops, a necessary step in the realization of automated disassembly systems. A study was performed that varied the lighting conditions, velocity, and number of passes the robot made over the laptop, as well as the network size of a YOLO-v5 neural network. The analysis reveals that specific combinations of operating parameters and neural network configurations can significantly improve detection accuracy. Specifically, the best results for the majority of laptops were obtained when the system ran at medium velocity (10 and 15 mm/s), with a light, and the neural network was run with an extra large network. Additionally, the results show that screw characteristics like the screw hole depth, the presence of a taper in the screw hole, screw hole location, and the color difference between the laptop cover and the screw color impact the system’s overall detection rate, with the most important factor being the depth of the screw. Full article
(This article belongs to the Special Issue Latest Research on Computer Vision and Image Processing)
Show Figures

Figure 1

24 pages, 7706 KiB  
Article
Computer Vision for Safety Management in the Steel Industry
by Roy Lan, Ibukun Awolusi and Jiannan Cai
AI 2024, 5(3), 1192-1215; https://doi.org/10.3390/ai5030058 (registering DOI) - 19 Jul 2024
Viewed by 131
Abstract
The complex nature of the steel manufacturing environment, characterized by different types of hazards from materials and large machinery, makes the need for objective and automated monitoring very critical to replace the traditional methods, which are manual and subjective. This study explores the [...] Read more.
The complex nature of the steel manufacturing environment, characterized by different types of hazards from materials and large machinery, makes the need for objective and automated monitoring very critical to replace the traditional methods, which are manual and subjective. This study explores the feasibility of implementing computer vision for safety management in steel manufacturing, with a case study implementation for automated hard hat detection. The research combines hazard characterization, technology assessment, and a pilot case study. First, a comprehensive review of steel manufacturing hazards was conducted, followed by the application of TOPSIS, a multi-criteria decision analysis method, to select a candidate computer vision system from eight commercially available systems. This pilot study evaluated YOLOv5m, YOLOv8m, and YOLOv9c models on 703 grayscale images from a steel mini-mill, assessing performance through precision, recall, F1-score, mAP, specificity, and AUC metrics. Results showed high overall accuracy in hard hat detection, with YOLOv9c slightly outperforming others, particularly in detecting safety violations. Challenges emerged in handling class imbalance and accurately identifying absent hard hats, especially given grayscale imagery limitations. Despite these challenges, this study affirms the feasibility of computer vision-based safety management in steel manufacturing, providing a foundation for future automated safety monitoring systems. Findings underscore the need for larger, diverse datasets and advanced techniques to address industry-specific complexities, paving the way for enhanced workplace safety in challenging industrial environments. Full article
(This article belongs to the Special Issue Artificial Intelligence-Based Image Processing and Computer Vision)
Show Figures

Figure 1

18 pages, 6988 KiB  
Article
SAW-YOLO: A Multi-Scale YOLO for Small Target Citrus Pests Detection
by Xiaojiang Wu, Jinzhe Liang, Yiyu Yang, Zhenghao Li, Xinyu Jia, Haibo Pu and Peng Zhu
Agronomy 2024, 14(7), 1571; https://doi.org/10.3390/agronomy14071571 (registering DOI) - 19 Jul 2024
Viewed by 109
Abstract
Citrus pests pose a major threat to both citrus yield and fruit quality. The early prevention of pests is essential for sustainable citrus cultivation, cost savings, and the reduction of environmental pollution. Despite the increasing application of deep learning techniques in agriculture, the [...] Read more.
Citrus pests pose a major threat to both citrus yield and fruit quality. The early prevention of pests is essential for sustainable citrus cultivation, cost savings, and the reduction of environmental pollution. Despite the increasing application of deep learning techniques in agriculture, the performance of existing models for small target detection of citrus pests is limited, mainly in terms of information bottlenecks that occur during the transfer of information. This hinders its effectiveness in fully automating the detection of citrus pests. In this study, a new approach was introduced to overcome these limitations. Firstly, a comprehensive large-scale dataset named IP-CitrusPests13 was developed, encompassing 13 distinct citrus pest categories. This dataset was amalgamated from IP102 and web crawlers, serving as a fundamental resource for precision-oriented pest detection tasks in citrus farming. Web crawlers can supplement information on various forms of pests and changes in pest size. Using this comprehensive dataset, we employed the SPD Module in the backbone network to preserve fine-grained information and prevent the model from losing important information as the depth increased. In addition, we introduced the AFFD Head detection module into the YOLOv8 architecture, which has two important functions that effectively integrate shallow and deep information to improve the learning ability of the model. Optimizing the bounding box loss function to WIoU v3 (Wise-IoU v3), which focuses on medium-quality anchor frames, sped up the convergence of the network. Experimental evaluation on a test set showed that the proposed SAW-YOLO (SPD Module, AFFD, WIoU v3) model achieved an average accuracy of 90.3%, which is 3.3% higher than the benchmark YOLOv8n model. Without any significant enlargement in the model size, state-of-the-art (SOTA) performance can be achieved in small target detection. To validate the robustness of the model against pests of various sizes, the SAW-YOLO model showed improved detection performance on all three scales of pests, significantly reducing the rate of missed detections. Our experimental results show that the SAW-YOLO model performs well in the detection of multiple pest classes in citrus orchards, helping to advance smart planting practices in the citrus industry. Full article
(This article belongs to the Special Issue AI, Sensors and Robotics for Smart Agriculture—2nd Edition)
Show Figures

Figure 1

17 pages, 5588 KiB  
Article
Detection of Liquid Retention on Pipette Tips in High-Throughput Liquid Handling Workstations Based on Improved YOLOv8 Algorithm with Attention Mechanism
by Yanpu Yin, Jiahui Lei and Wei Tao
Electronics 2024, 13(14), 2836; https://doi.org/10.3390/electronics13142836 (registering DOI) - 18 Jul 2024
Viewed by 301
Abstract
High-throughput liquid handling workstations are required to process large numbers of test samples in the fields of life sciences and medicine. Liquid retention and droplets hanging in the pipette tips can lead to cross-contamination of samples and reagents and inaccurate experimental results. Traditional [...] Read more.
High-throughput liquid handling workstations are required to process large numbers of test samples in the fields of life sciences and medicine. Liquid retention and droplets hanging in the pipette tips can lead to cross-contamination of samples and reagents and inaccurate experimental results. Traditional methods for detecting liquid retention have low precision and poor real-time performance. This paper proposes an improved YOLOv8 (You Only Look Once version 8) object detection algorithm to address the challenges posed by different liquid sizes and colors, complex situation of test tube racks and multiple samples in the background, and poor global image structure understanding in pipette tip liquid retention detection. A global context (GC) attention mechanism module is introduced into the backbone network and the cross-stage partial feature fusion (C2f) module to better focus on target features. To enhance the ability to effectively combine and process different types of data inputs and background information, a Large Kernel Selection (LKS) module is also introduced into the backbone network. Additionally, the neck network is redesigned to incorporate the Simple Attention (SimAM) mechanism module, generating attention weights and improving overall performance. We evaluated the algorithm using a self-built dataset of pipette tips. Compared to the original YOLOv8 model, the improved algorithm increased [email protected] (mean average precision), F1 score, and precision by 1.7%, 2%, and 1.7%, respectively. The improved YOLOv8 algorithm can enhance the detection capability of liquid-retaining pipette tips, and prevent cross-contamination from affecting the results of sample solution experiments. It provides a detection basis for subsequent automatic processing of solution for liquid retention. Full article
(This article belongs to the Special Issue Advances in Computer Vision and Deep Learning and Its Applications)
Show Figures

Figure 1

21 pages, 13465 KiB  
Article
Vision-Based Anti-UAV Detection Based on YOLOv7-GS in Complex Backgrounds
by Chunjuan Bo, Yuntao Wei, Xiujia Wang, Zhan Shi and Ying Xiao
Drones 2024, 8(7), 331; https://doi.org/10.3390/drones8070331 - 18 Jul 2024
Viewed by 169
Abstract
Unauthorized unmanned aerial vehicles (UAVs) pose threats to public safety and individual privacy. Traditional object-detection approaches often fall short during their application in anti-UAV technologies. To address this issue, we propose the YOLOv7-GS model, which is designed specifically for the identification of small [...] Read more.
Unauthorized unmanned aerial vehicles (UAVs) pose threats to public safety and individual privacy. Traditional object-detection approaches often fall short during their application in anti-UAV technologies. To address this issue, we propose the YOLOv7-GS model, which is designed specifically for the identification of small UAVs in complex and low-altitude environments. This research primarily aims to improve the model’s detection capabilities for small UAVs in complex backgrounds. Enhancements were applied to the YOLOv7-tiny model, including adjustments to the sizes of prior boxes, incorporation of the InceptionNeXt module at the end of the neck section, and introduction of the SPPFCSPC-SR and Get-and-Send modules. These modifications aid in the preservation of details about small UAVs and heighten the model’s focus on them. The YOLOv7-GS model achieves commendable results on the DUT Anti-UAV and the Amateur Unmanned Air Vehicle Detection datasets and performs to be competitive against other mainstream algorithms. Full article
Show Figures

Figure 1

33 pages, 26737 KiB  
Article
EcoDetect-YOLO: A Lightweight, High-Generalization Methodology for Real-Time Detection of Domestic Waste Exposure in Intricate Environmental Landscapes
by Shenlin Liu, Ruihan Chen, Minhua Ye, Jiawei Luo, Derong Yang and Ming Dai
Sensors 2024, 24(14), 4666; https://doi.org/10.3390/s24144666 - 18 Jul 2024
Viewed by 222
Abstract
In response to the challenges of accurate identification and localization of garbage in intricate urban street environments, this paper proposes EcoDetect-YOLO, a garbage exposure detection algorithm based on the YOLOv5s framework, utilizing an intricate environment waste exposure detection dataset constructed in this study. [...] Read more.
In response to the challenges of accurate identification and localization of garbage in intricate urban street environments, this paper proposes EcoDetect-YOLO, a garbage exposure detection algorithm based on the YOLOv5s framework, utilizing an intricate environment waste exposure detection dataset constructed in this study. Initially, a convolutional block attention module (CBAM) is integrated between the second level of the feature pyramid etwork (P2) and the third level of the feature pyramid network (P3) layers to optimize the extraction of relevant garbage features while mitigating background noise. Subsequently, a P2 small-target detection head enhances the model’s efficacy in identifying small garbage targets. Lastly, a bidirectional feature pyramid network (BiFPN) is introduced to strengthen the model’s capability for deep feature fusion. Experimental results demonstrate EcoDetect-YOLO’s adaptability to urban environments and its superior small-target detection capabilities, effectively recognizing nine types of garbage, such as paper and plastic trash. Compared to the baseline YOLOv5s model, EcoDetect-YOLO achieved a 4.7% increase in mAP0.5, reaching 58.1%, with a compact model size of 15.7 MB and an FPS of 39.36. Notably, even in the presence of strong noise, the model maintained a mAP0.5 exceeding 50%, underscoring its robustness. In summary, EcoDetect-YOLO, as proposed in this paper, boasts high precision, efficiency, and compactness, rendering it suitable for deployment on mobile devices for real-time detection and management of urban garbage exposure, thereby advancing urban automation governance and digital economic development. Full article
(This article belongs to the Section Physical Sensors)
Show Figures

Figure 1

16 pages, 9223 KiB  
Article
NATCA YOLO-Based Small Object Detection for Aerial Images
by Yicheng Zhu, Zhenhua Ai, Jinqiang Yan, Silong Li, Guowei Yang and Teng Yu
Information 2024, 15(7), 414; https://doi.org/10.3390/info15070414 - 18 Jul 2024
Viewed by 209
Abstract
The object detection model in UAV aerial image scenes faces challenges such as significant scale changes of certain objects and the presence of complex backgrounds. This paper aims to address the detection of small objects in aerial images using NATCA (neighborhood attention Transformer [...] Read more.
The object detection model in UAV aerial image scenes faces challenges such as significant scale changes of certain objects and the presence of complex backgrounds. This paper aims to address the detection of small objects in aerial images using NATCA (neighborhood attention Transformer coordinate attention) YOLO. Specifically, the feature extraction network incorporates a neighborhood attention transformer (NAT) into the last layer to capture global context information and extract diverse features. Additionally, the feature fusion network (Neck) incorporates a coordinate attention (CA) module to capture channel information and longer-range positional information. Furthermore, the activation function in the original convolutional block is replaced with Meta-ACON. The NAT serves as the prediction layer in the new network, which is evaluated using the VisDrone2019-DET object detection dataset as a benchmark, and tested on the VisDrone2019-DET-test-dev dataset. To assess the performance of the NATCA YOLO model in detecting small objects in aerial images, other detection networks, such as Faster R-CNN, RetinaNet, and SSD, are employed for comparison on the test set. The results demonstrate that the NATCA YOLO detection achieves an average accuracy of 42%, which is a 2.9% improvement compared to the state-of-the-art detection network TPH-YOLOv5. Full article
Show Figures

Figure 1

22 pages, 23929 KiB  
Article
FireYOLO-Lite: Lightweight Forest Fire Detection Network with Wide-Field Multi-Scale Attention Mechanism
by Sha Sheng, Zhengyin Liang, Wenxing Xu, Yong Wang and Jiangdan Su
Forests 2024, 15(7), 1244; https://doi.org/10.3390/f15071244 - 17 Jul 2024
Viewed by 214
Abstract
A lightweight forest fire detection model based on YOLOv8 is proposed in this paper in response to the problems existing in traditional sensors for forest fire detection. The performance of traditional sensors is easily constrained by hardware computing power, and their adaptability in [...] Read more.
A lightweight forest fire detection model based on YOLOv8 is proposed in this paper in response to the problems existing in traditional sensors for forest fire detection. The performance of traditional sensors is easily constrained by hardware computing power, and their adaptability in different environments needs improvement. To balance the accuracy and speed of fire detection, the GhostNetV2 lightweight network is adopted to replace the backbone network for feature extraction of YOLOv8. The Ghost module is utilized to replace traditional convolution operations, conducting feature extraction independently in different dimensional channels, significantly reducing the complexity of the model while maintaining excellent performance. Additionally, an improved CPDCA channel priority attention mechanism is proposed, which extracts spatial features through dilated convolution, thereby reducing computational overhead and enabling the model to focus more on fire targets, achieving more accurate detection. In response to the problem of small targets in fire detection, the Inner IoU loss function is introduced. By adjusting the size of the auxiliary bounding boxes, this function effectively enhances the convergence effect of small target detection, further reducing missed detections, and improving overall detection accuracy. Experimental results indicate that, compared with traditional methods, the algorithm proposed in this paper significantly improves the average precision and FPS of fire detection while maintaining a smaller model size. Through experimental analysis, compared with YOLOv3-tiny, the average precision increased by 5.9% and the frame rate reached 285.3 FPS when the model size was only 4.9 M; compared with Shufflenet, the average precision increased by 2.9%, and the inference speed tripled. Additionally, the algorithm effectively addresses false positives, such as cloud and reflective light, further enhancing the detection of small targets and reducing missed detections. Full article
(This article belongs to the Section Natural Hazards and Risk Management)
Show Figures

Figure 1

16 pages, 9904 KiB  
Article
Improved Chinese Giant Salamander Parental Care Behavior Detection Based on YOLOv8
by Zhihao Li, Shouliang Luo, Jing Xiang, Yuanqiong Chen and Qinghua Luo
Animals 2024, 14(14), 2089; https://doi.org/10.3390/ani14142089 - 17 Jul 2024
Viewed by 265
Abstract
Optimizing the breeding techniques and increasing the hatching rate of Andrias davidianus offspring necessitates a thorough understanding of its parental care behaviors. However, A. davidianus’ nocturnal and cave-dwelling tendencies pose significant challenges for direct observation. To address this problem, this study constructed [...] Read more.
Optimizing the breeding techniques and increasing the hatching rate of Andrias davidianus offspring necessitates a thorough understanding of its parental care behaviors. However, A. davidianus’ nocturnal and cave-dwelling tendencies pose significant challenges for direct observation. To address this problem, this study constructed a dataset for the parental care behavior of A. davidianus, applied the target detection method to this behavior for the first time, and proposed a detection model for A. davidianus’ parental care behavior based on the YOLOv8s algorithm. Firstly, a multi-scale feature fusion convolution (MSConv) is proposed and combined with a C2f module, which significantly enhances the feature extraction capability of the model. Secondly, the large separable kernel attention is introduced into the spatial pyramid pooling fast (SPPF) layer to effectively reduce the interference factors in the complex environment. Thirdly, to address the problem of low quality of captured images, Wise-IoU (WIoU) is used to replace CIoU in the original YOLOv8 to optimize the loss function and improve the model’s robustness. The experimental results show that the model achieves 85.7% in the mAP50-95, surpassing the YOLOv8s model by 2.1%. Compared with other mainstream models, the overall performance of our model is much better and can effectively detect the parental care behavior of A. davidianus. Our research method not only offers a reference for the behavior recognition of A. davidianus and other amphibians but also provides a new strategy for the smart breeding of A. davidianus. Full article
Show Figures

Figure 1

Back to TopTop