Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (170)

Search Parameters:
Keywords = BiFPN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
33 pages, 26737 KiB  
Article
EcoDetect-YOLO: A Lightweight, High-Generalization Methodology for Real-Time Detection of Domestic Waste Exposure in Intricate Environmental Landscapes
by Shenlin Liu, Ruihan Chen, Minhua Ye, Jiawei Luo, Derong Yang and Ming Dai
Sensors 2024, 24(14), 4666; https://doi.org/10.3390/s24144666 - 18 Jul 2024
Viewed by 144
Abstract
In response to the challenges of accurate identification and localization of garbage in intricate urban street environments, this paper proposes EcoDetect-YOLO, a garbage exposure detection algorithm based on the YOLOv5s framework, utilizing an intricate environment waste exposure detection dataset constructed in this study. [...] Read more.
In response to the challenges of accurate identification and localization of garbage in intricate urban street environments, this paper proposes EcoDetect-YOLO, a garbage exposure detection algorithm based on the YOLOv5s framework, utilizing an intricate environment waste exposure detection dataset constructed in this study. Initially, a convolutional block attention module (CBAM) is integrated between the second level of the feature pyramid etwork (P2) and the third level of the feature pyramid network (P3) layers to optimize the extraction of relevant garbage features while mitigating background noise. Subsequently, a P2 small-target detection head enhances the model’s efficacy in identifying small garbage targets. Lastly, a bidirectional feature pyramid network (BiFPN) is introduced to strengthen the model’s capability for deep feature fusion. Experimental results demonstrate EcoDetect-YOLO’s adaptability to urban environments and its superior small-target detection capabilities, effectively recognizing nine types of garbage, such as paper and plastic trash. Compared to the baseline YOLOv5s model, EcoDetect-YOLO achieved a 4.7% increase in mAP0.5, reaching 58.1%, with a compact model size of 15.7 MB and an FPS of 39.36. Notably, even in the presence of strong noise, the model maintained a mAP0.5 exceeding 50%, underscoring its robustness. In summary, EcoDetect-YOLO, as proposed in this paper, boasts high precision, efficiency, and compactness, rendering it suitable for deployment on mobile devices for real-time detection and management of urban garbage exposure, thereby advancing urban automation governance and digital economic development. Full article
(This article belongs to the Section Physical Sensors)
Show Figures

Figure 1

22 pages, 9246 KiB  
Article
DST-DETR: Image Dehazing RT-DETR for Safety Helmet Detection in Foggy Weather
by Ziyuan Liu, Chunxia Sun and Xiaopeng Wang
Sensors 2024, 24(14), 4628; https://doi.org/10.3390/s24144628 - 17 Jul 2024
Viewed by 161
Abstract
In foggy weather, outdoor safety helmet detection often suffers from low visibility and unclear objects, hindering optimal detector performance. Moreover, safety helmets typically appear as small objects at construction sites, prone to occlusion and difficult to distinguish from complex backgrounds, further exacerbating the [...] Read more.
In foggy weather, outdoor safety helmet detection often suffers from low visibility and unclear objects, hindering optimal detector performance. Moreover, safety helmets typically appear as small objects at construction sites, prone to occlusion and difficult to distinguish from complex backgrounds, further exacerbating the detection challenge. Therefore, the real-time and precise detection of safety helmet usage among construction personnel, particularly in adverse weather conditions such as foggy weather, poses a significant challenge. To address this issue, this paper proposes the DST-DETR, a framework for foggy weather safety helmet detection. The DST-DETR framework comprises a dehazing module, PAOD-Net, and an object detection module, ST-DETR, for joint dehazing and detection. Initially, foggy images are restored within PAOD-Net, enhancing the AOD-Net model by introducing a novel convolutional module, PfConv, guided by the parameter-free average attention module (PfAAM). This module enables more focused attention on crucial features in lightweight models, therefore enhancing performance. Subsequently, the MS-SSIM + 2 loss function is employed to bolster the model’s robustness, making it adaptable to scenes with intricate backgrounds and variable fog densities. Next, within the object detection module, the ST-DETR model is designed to address small objects. By refining the RT-DETR model, its capability to detect small objects in low-quality images is enhanced. The core of this approach lies in utilizing the variant ResNet-18 as the backbone to make the network lightweight without sacrificing accuracy, followed by effectively integrating the small-object layer into the improved BiFPN neck structure, resulting in CCFF-BiFPN-P2. Various experiments were conducted to qualitatively and quantitatively compare our method with several state-of-the-art approaches, demonstrating its superiority. The results validate that the DST-DETR algorithm is better suited for foggy safety helmet detection tasks in construction scenarios. Full article
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)
Show Figures

Figure 1

18 pages, 5924 KiB  
Article
Multi-Scale Marine Object Detection in Side-Scan Sonar Images Based on BES-YOLO
by Quanhong Ma, Shaohua Jin, Gang Bian and Yang Cui
Sensors 2024, 24(14), 4428; https://doi.org/10.3390/s24144428 - 9 Jul 2024
Viewed by 360
Abstract
Aiming at the problem of low accuracy of multi-scale seafloor target detection in side-scan sonar images with high noise and complex background texture, a model for multi-scale target detection using the BES-YOLO network is proposed. First, an efficient multi-scale attention (EMA) mechanism is [...] Read more.
Aiming at the problem of low accuracy of multi-scale seafloor target detection in side-scan sonar images with high noise and complex background texture, a model for multi-scale target detection using the BES-YOLO network is proposed. First, an efficient multi-scale attention (EMA) mechanism is used in the backbone of the YOLOv8 network, and a bi-directional feature pyramid network (Bifpn) is introduced to merge the information of different scales, finally, a Shape_IoU loss function is introduced to continuously optimize the model and improve its accuracy. Before training, the dataset is preprocessed using 2D discrete wavelet decomposition and reconstruction to enhance the robustness of the network. The experimental results show that 92.4% of the mean average accuracy at IoU of 0.5 ([email protected]) and 67.7% of the mean average accuracy at IoU of 0.5 to 0.95 ([email protected]:0.95) are achieved using the BES-YOLO network, which is an increase of 5.3% and 4.4% compared to the YOLOv8n model. The research results can effectively improve the detection accuracy and efficiency of multi-scale targets in side-scan sonar images, which can be applied to AUVs and other underwater platforms to implement intelligent detection of undersea targets. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

15 pages, 1293 KiB  
Article
An Improved Lightweight YOLOv5s-Based Method for Detecting Electric Bicycles in Elevators
by Ziyuan Zhang, Xianyu Yang and Chengyu Wu
Electronics 2024, 13(13), 2660; https://doi.org/10.3390/electronics13132660 - 7 Jul 2024
Viewed by 372
Abstract
The increase in fire accidents caused by indoor charging of electric bicycles has raised concerns among people. Monitoring EBs in elevators is challenging, and the current object detection method is a variant of YOLOv5, which faces problems with calculating the load and detection [...] Read more.
The increase in fire accidents caused by indoor charging of electric bicycles has raised concerns among people. Monitoring EBs in elevators is challenging, and the current object detection method is a variant of YOLOv5, which faces problems with calculating the load and detection rate. To address this issue, this paper presents an improved lightweight method based on YOLOv5s to detect EBs in elevators. This method introduces the MobileNetV2 module to achieve the lightweight performance of the model. By introducing the CBAM attention mechanism and the Bidirectional Feature Pyramid Network (BiFPN) into the YOLOv5s neck network, the detection precision is improved. In order to better verify that the model can be deployed at the edge of an elevator, this article deploys it using the Raspberry Pi 4B embedded development board and connects it to a buzzer for application verification. The experimental results demonstrate that the model parameters of EBs are reduced by 58.4%, the computational complexity is reduced by 50.6%, the detection precision reaches 95.9%, and real-time detection of electric vehicles in elevators is achieved. Full article
Show Figures

Figure 1

17 pages, 20371 KiB  
Article
YOLOv8 Model for Weed Detection in Wheat Fields Based on a Visual Converter and Multi-Scale Feature Fusion
by Yinzeng Liu, Fandi Zeng, Hongwei Diao, Junke Zhu, Dong Ji, Xijie Liao and Zhihuan Zhao
Sensors 2024, 24(13), 4379; https://doi.org/10.3390/s24134379 - 5 Jul 2024
Viewed by 362
Abstract
Accurate weed detection is essential for the precise control of weeds in wheat fields, but weeds and wheat are sheltered from each other, and there is no clear size specification, making it difficult to accurately detect weeds in wheat. To achieve the precise [...] Read more.
Accurate weed detection is essential for the precise control of weeds in wheat fields, but weeds and wheat are sheltered from each other, and there is no clear size specification, making it difficult to accurately detect weeds in wheat. To achieve the precise identification of weeds, wheat weed datasets were constructed, and a wheat field weed detection model, YOLOv8-MBM, based on improved YOLOv8s, was proposed. In this study, a lightweight visual converter (MobileViTv3) was introduced into the C2f module to enhance the detection accuracy of the model by integrating input, local (CNN), and global (ViT) features. Secondly, a bidirectional feature pyramid network (BiFPN) was introduced to enhance the performance of multi-scale feature fusion. Furthermore, to address the weak generalization and slow convergence speed of the CIoU loss function for detection tasks, the bounding box regression loss function (MPDIOU) was used instead of the CIoU loss function to improve the convergence speed of the model and further enhance the detection performance. Finally, the model performance was tested on the wheat weed datasets. The experiments show that the YOLOv8-MBM proposed in this paper is superior to Fast R-CNN, YOLOv3, YOLOv4-tiny, YOLOv5s, YOLOv7, YOLOv9, and other mainstream models in regards to detection performance. The accuracy of the improved model reaches 92.7%. Compared with the original YOLOv8s model, the precision, recall, mAP1, and mAP2 are increased by 10.6%, 8.9%, 9.7%, and 9.3%, respectively. In summary, the YOLOv8-MBM model successfully meets the requirements for accurate weed detection in wheat fields. Full article
(This article belongs to the Special Issue Sensor and AI Technologies in Intelligent Agriculture: 2nd Edition)
Show Figures

Figure 1

14 pages, 4596 KiB  
Article
BiFPN-KPointNet-CBAM: Application of 3D Point Cloud Technology Based on Deep Learning in Measuring Vegetation
by Qihuanghua Liu, Jianmin Jiang, Jingyi Hu, Songyu Zhong and Fang Zou
Electronics 2024, 13(13), 2577; https://doi.org/10.3390/electronics13132577 - 30 Jun 2024
Viewed by 431
Abstract
The results of traditional vegetation-measuring methods are mostly two-dimensional data, which can only convey limited information. The greening situation of many cities or regions in the world cannot be fully assessed by these results. In this regard, this paper proposes the use of [...] Read more.
The results of traditional vegetation-measuring methods are mostly two-dimensional data, which can only convey limited information. The greening situation of many cities or regions in the world cannot be fully assessed by these results. In this regard, this paper proposes the use of the air–ground integrated point cloud data acquisition mode for measuring vegetation. This mode combines a backpack-mounted laser scanning system, a vehicle-mounted laser scanning system, and UAV tilt photography technology to collect greening data in a comprehensive park and along a municipal road in Guangzhou, China. To classify the collected greening data, we propose the BiFPN-KPointNet-CBAM model, which was derived from PointNet. The model was introduced to analyze the distribution of green plants in study areas. The experimental findings indicate that our model achieved a notable enhancement in the overall accuracy by approximately 8% compared with other state-of-the-art models. Compared with the traditional greening survey method, this method obtained three-dimensional and more accurate greening data, and thus, provides higher quality greening data for urban managers. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

25 pages, 3845 KiB  
Article
Bud-YOLOv8s: A Potato Bud-Eye-Detection Algorithm Based on Improved YOLOv8s
by Wenlong Liu, Zhao Li, Shaoshuang Zhang, Ting Qin and Jiaqi Zhao
Electronics 2024, 13(13), 2541; https://doi.org/10.3390/electronics13132541 - 28 Jun 2024
Viewed by 556
Abstract
The key to intelligent seed potato cutting technology lies in the accurate and rapid identification of potato bud eyes. Existing detection algorithms suffer from low recognition accuracy and high model complexity, resulting in an increased miss rate. To address these issues, this study [...] Read more.
The key to intelligent seed potato cutting technology lies in the accurate and rapid identification of potato bud eyes. Existing detection algorithms suffer from low recognition accuracy and high model complexity, resulting in an increased miss rate. To address these issues, this study proposes a potato bud-eye-detection algorithm based on an improved YOLOv8s. First, by integrating the Faster Neural Network (FasterNet) with the Efficient Multi-scale Attention (EMA) module, a novel Faster Block-EMA network structure is designed to replace the bottleneck components within the C2f module of YOLOv8s. This enhancement improves the model’s feature-extraction capability and computational efficiency for bud detection. Second, this study introduces a weighted bidirectional feature pyramid network (BiFPN) to optimize the neck network, achieving multi-scale fusion of potato bud eye features while significantly reducing the model’s parameters, computation, and size due to its flexible network topology. Finally, the Efficient Intersection over Union (EIoU) loss function is employed to optimize the bounding box regression process, further enhancing the model’s localization capability. The experimental results show that the improved model achieves a mean average precision ([email protected]) of 98.1% with a model size of only 11.1 MB. Compared to the baseline model, the [email protected] and [email protected]:0.95 were improved by 3.1% and 4.5%, respectively, while the model’s parameters, size, and computation were reduced by 49.1%, 48.1%, and 31.1%, respectively. Additionally, compared to the YOLOv3, YOLOv5s, YOLOv6s, YOLOv7-tiny, and YOLOv8m algorithms, the [email protected] was improved by 4.6%, 3.7%, 5.6%, 5.2%, and 3.3%, respectively. Therefore, the proposed algorithm not only significantly enhances the detection accuracy, but also greatly reduces the model complexity, providing essential technical support for the application and deployment of intelligent potato cutting technology. Full article
Show Figures

Figure 1

14 pages, 4284 KiB  
Article
Leaf Segmentation Using Modified YOLOv8-Seg Models
by Peng Wang, Hong Deng, Jiaxu Guo, Siqi Ji, Dan Meng, Jun Bao and Peng Zuo
Life 2024, 14(6), 780; https://doi.org/10.3390/life14060780 - 20 Jun 2024
Viewed by 551
Abstract
Computer-vision-based plant leaf segmentation technology is of great significance for plant classification, monitoring of plant growth, precision agriculture, and other scientific research. In this paper, the YOLOv8-seg model was used for the automated segmentation of individual leaves in images. In order to improve [...] Read more.
Computer-vision-based plant leaf segmentation technology is of great significance for plant classification, monitoring of plant growth, precision agriculture, and other scientific research. In this paper, the YOLOv8-seg model was used for the automated segmentation of individual leaves in images. In order to improve the segmentation performance, we further introduced a Ghost module and a Bidirectional Feature Pyramid Network (BiFPN) module into the standard Yolov8 model and proposed two modified versions. The Ghost module can generate several intrinsic feature maps with cheap transformation operations, and the BiFPN module can fuse multi-scale features to improve the segmentation performance of small leaves. The experiment results show that Yolov8 performs well in the leaf segmentation task, and the Ghost module and BiFPN module can further improve the performance. Our proposed approach achieves a 86.4% leaf segmentation score (best Dice) over all five test datasets of the Computer Vision Problems in Plant Phenotyping (CVPPP) Leaf Segmentation Challenge, outperforming other reported approaches. Full article
(This article belongs to the Section Plant Science)
Show Figures

Figure 1

23 pages, 21902 KiB  
Article
WH-DETR: An Efficient Network Architecture for Wheat Spike Detection in Complex Backgrounds
by Zhenlin Yang, Wanhong Yang, Jizheng Yi and Rong Liu
Agriculture 2024, 14(6), 961; https://doi.org/10.3390/agriculture14060961 - 19 Jun 2024
Viewed by 456
Abstract
Wheat spike detection is crucial for estimating wheat yields and has a significant impact on the modernization of wheat cultivation and the advancement of precision agriculture. This study explores the application of the DETR (Detection Transformer) architecture in wheat spike detection, introducing a [...] Read more.
Wheat spike detection is crucial for estimating wheat yields and has a significant impact on the modernization of wheat cultivation and the advancement of precision agriculture. This study explores the application of the DETR (Detection Transformer) architecture in wheat spike detection, introducing a new perspective to this task. We propose a high-precision end-to-end network named WH-DETR, which is based on an enhanced RT-DETR architecture. Initially, we employ data augmentation techniques such as image rotation, scaling, and random occlusion on the GWHD2021 dataset to improve the model’s generalization across various scenarios. A lightweight feature pyramid, GS-BiFPN, is implemented in the network’s neck section to effectively extract the multi-scale features of wheat spikes in complex environments, such as those with occlusions, overlaps, and extreme lighting conditions. Additionally, the introduction of GSConv enhances the network precision while reducing the computational costs, thereby controlling the detection speed. Furthermore, the EIoU metric is integrated into the loss function, refined to better focus on partially occluded or overlapping spikes. The testing results on the dataset demonstrate that this method achieves an Average Precision (AP) of 95.7%, surpassing current state-of-the-art object detection methods in both precision and speed. These findings confirm that our approach more closely meets the practical requirements for wheat spike detection compared to existing methods. Full article
(This article belongs to the Special Issue Computer Vision and Artificial Intelligence in Agriculture)
Show Figures

Figure 1

17 pages, 9027 KiB  
Article
DV3-IBi_YOLOv5s: A Lightweight Backbone Network and Multiscale Neck Network Vehicle Detection Algorithm
by Liu Wang, Lijuan Shi, Jian Zhao, Chen Yang, Haixia Li, Yaodong Jia and Haiyan Wang
Sensors 2024, 24(12), 3791; https://doi.org/10.3390/s24123791 - 11 Jun 2024
Viewed by 372
Abstract
Vehicle detection is a research direction in the field of target detection and is widely used in intelligent transportation, automatic driving, urban planning, and other fields. To balance the high-speed advantage of lightweight networks and the high-precision advantage of multiscale networks, a vehicle [...] Read more.
Vehicle detection is a research direction in the field of target detection and is widely used in intelligent transportation, automatic driving, urban planning, and other fields. To balance the high-speed advantage of lightweight networks and the high-precision advantage of multiscale networks, a vehicle detection algorithm based on a lightweight backbone network and a multiscale neck network is proposed. The mobile NetV3 lightweight network based on deep separable convolution is used as the backbone network to improve the speed of vehicle detection. The icbam attention mechanism module is used to strengthen the processing of the vehicle feature information detected by the backbone network to enrich the input information of the neck network. The bifpn and icbam attention mechanism modules are integrated into the neck network to improve the detection accuracy of vehicles of different sizes and categories. A vehicle detection experiment on the Ua-Detrac dataset verifies that the proposed algorithm can effectively balance vehicle detection accuracy and speed. The detection accuracy is 71.19%, the number of parameters is 3.8 MB, and the detection speed is 120.02 fps, which meets the actual requirements of the parameter quantity, detection speed, and accuracy of the vehicle detection algorithm embedded in the mobile device. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

17 pages, 10207 KiB  
Article
Improved YOLOv8-Seg Based on Multiscale Feature Fusion and Deformable Convolution for Weed Precision Segmentation
by Zhuxi Lyu, Anjiang Lu and Yinglong Ma
Appl. Sci. 2024, 14(12), 5002; https://doi.org/10.3390/app14125002 - 7 Jun 2024
Viewed by 692
Abstract
Laser-targeted weeding methods further enhance the sustainable development of green agriculture, with one key technology being the improvement of weed localization accuracy. Here, we propose an improved YOLOv8 instance segmentation based on bidirectional feature fusion and deformable convolution (BFFDC-YOLOv8-seg) to address the challenges [...] Read more.
Laser-targeted weeding methods further enhance the sustainable development of green agriculture, with one key technology being the improvement of weed localization accuracy. Here, we propose an improved YOLOv8 instance segmentation based on bidirectional feature fusion and deformable convolution (BFFDC-YOLOv8-seg) to address the challenges of insufficient weed localization accuracy in complex environments with resource-limited laser weeding devices. Initially, by training on extensive datasets of plant images, the most appropriate model scale and training weights are determined, facilitating the development of a lightweight network. Subsequently, the introduction of the Bidirectional Feature Pyramid Network (BiFPN) during feature fusion effectively prevents the omission of weeds. Lastly, the use of Dynamic Snake Convolution (DSConv) to replace some convolutional kernels enhances flexibility, benefiting the segmentation of weeds with elongated stems and irregular edges. Experimental results indicate that the BFFDC-YOLOv8-seg model achieves a 4.9% increase in precision, an 8.1% increase in recall rate, and a 2.8% increase in mAP50 value to 98.8% on a vegetable weed dataset compared to the original model. It also shows improved mAP50 over other typical segmentation models such as Mask R-CNN, YOLOv5-seg, and YOLOv7-seg by 10.8%, 13.4%, and 1.8%, respectively. Furthermore, the model achieves a detection speed of 24.8 FPS on the Jetson Orin nano standalone device, with a model size of 6.8 MB that balances between size and accuracy. The model meets the requirements for real-time precise weed segmentation, and is suitable for complex vegetable field environments and resource-limited laser weeding devices. Full article
(This article belongs to the Section Agricultural Science and Technology)
Show Figures

Figure 1

16 pages, 5742 KiB  
Article
RSDNet: A New Multiscale Rail Surface Defect Detection Model
by Jingyi Du, Ruibo Zhang, Rui Gao, Lei Nan and Yifan Bao
Sensors 2024, 24(11), 3579; https://doi.org/10.3390/s24113579 - 1 Jun 2024
Viewed by 390
Abstract
The rapid and accurate identification of rail surface defects is critical to the maintenance and operational safety of the rail. For the problems of large-scale differences in rail surface defects and many small-scale defects, this paper proposes a rail surface defect detection algorithm, [...] Read more.
The rapid and accurate identification of rail surface defects is critical to the maintenance and operational safety of the rail. For the problems of large-scale differences in rail surface defects and many small-scale defects, this paper proposes a rail surface defect detection algorithm, RSDNet (Rail Surface Defect Detection Net), with YOLOv8n as the baseline model. Firstly, the CDConv (Cascade Dilated Convolution) module is designed to realize multi-scale convolution by cascading the cavity convolution with different cavity rates. The CDConv is embedded into the backbone network to gather earlier defect local characteristics and contextual data. Secondly, the feature fusion method of Head is optimized based on BiFPN (Bi-directional Feature Pyramids Network) to fuse more layers of feature information and improve the utilization of original information. Finally, the EMA (Efficient Multi-Scale Attention) attention module is introduced to enhance the network’s attention to defect information. The experiments are conducted on the RSDDs dataset, and the experimental results show that the RSDNet algorithm achieves a mAP of 95.4% for rail surface defect detection, which is 4.6% higher than the original YOLOv8n. This study provides an effective technical means for rail surface defect detection that has certain engineering applications. Full article
Show Figures

Figure 1

20 pages, 5234 KiB  
Article
SDD-YOLO: A Lightweight, High-Generalization Methodology for Real-Time Detection of Strip Surface Defects
by Yueyang Wu, Ruihan Chen, Zhi Li, Minhua Ye and Ming Dai
Metals 2024, 14(6), 650; https://doi.org/10.3390/met14060650 - 30 May 2024
Viewed by 480
Abstract
Flat-rolled steel sheets are one of the major products of the metal industry. Strip steel’s production quality is crucial for the economic and safety aspects of humanity. Addressing the challenges of identifying the surface defects of strip steel in real production environments and [...] Read more.
Flat-rolled steel sheets are one of the major products of the metal industry. Strip steel’s production quality is crucial for the economic and safety aspects of humanity. Addressing the challenges of identifying the surface defects of strip steel in real production environments and low detection efficiency, this study presents an approach for strip defect detection based on YOLOv5s, termed SDD-YOLO. Initially, this study designs the Convolution-GhostNet Hybrid module (CGH) and Multi-Convolution Feature Fusion block (MCFF), effectively reducing computational complexity and enhancing feature extraction efficiency. Subsequently, CARAFE is employed to replace bilinear interpolation upsampling to improve image feature utilization; finally, the Bidirectional Feature Pyramid Network (BiFPN) is introduced to enhance the model’s adaptability to targets of different scales. Experimental results demonstrate that, compared to the baseline YOLOv5s, this method achieves a 6.3% increase in mAP50, reaching 76.1% on the Northeastern University Surface Defect Database for Detection (NEU-DET), with parameters and FLOPs of only 3.4MB and 6.4G, respectively, and FPS reaching 121, effectively identifying six types of defects such as Crazing and Inclusion. Furthermore, under the conditions of strong exposure, insufficient brightness, and the addition of Gaussian noise, the model’s mAP50 still exceeds 70%, demonstrating the model’s strong robustness. In conclusion, the proposed SDD-YOLO in this study features high accuracy, efficiency, and lightweight characteristics, making it applicable in actual production to enhance strip steel production quality and efficiency. Full article
(This article belongs to the Special Issue Machine Learning Models in Metals)
Show Figures

Figure 1

17 pages, 4647 KiB  
Article
Fine Segmentation of Chinese Character Strokes Based on Coordinate Awareness and Enhanced BiFPN
by Henghui Mo and Linjing Wei
Sensors 2024, 24(11), 3480; https://doi.org/10.3390/s24113480 - 28 May 2024
Viewed by 485
Abstract
Considering the complex structure of Chinese characters, particularly the connections and intersections between strokes, there are challenges in low accuracy of Chinese character stroke extraction and recognition, as well as unclear segmentation. This study builds upon the YOLOv8n-seg model to propose the YOLOv8n-seg-CAA-BiFPN [...] Read more.
Considering the complex structure of Chinese characters, particularly the connections and intersections between strokes, there are challenges in low accuracy of Chinese character stroke extraction and recognition, as well as unclear segmentation. This study builds upon the YOLOv8n-seg model to propose the YOLOv8n-seg-CAA-BiFPN Chinese character stroke fine segmentation model. The proposed Coordinate-Aware Attention mechanism (CAA) divides the backbone network input feature map into four parts, applying different weights for horizontal, vertical, and channel attention to compute and fuse key information, thus capturing the contextual regularity of closely arranged stroke positions. The network’s neck integrates an enhanced weighted bi-directional feature pyramid network (BiFPN), enhancing the fusion effect for features of strokes of various sizes. The Shape-IoU loss function is adopted in place of the traditional CIoU loss function, focusing on the shape and scale of stroke bounding boxes to optimize the bounding box regression process. Finally, the Grad-CAM++ technique is used to generate heatmaps of segmentation predictions, facilitating the visualization of effective features and a deeper understanding of the model’s focus areas. Trained and tested on the public Chinese character stroke datasets CCSE-Kai and CCSE-HW, the model achieves an average accuracy of 84.71%, an average recall rate of 83.65%, and a mean average precision of 80.11%. Compared to the original YOLOv8n-seg and existing mainstream segmentation models like SegFormer, BiSeNetV2, and Mask R-CNN, the average accuracy improved by 3.50%, 4.35%, 10.56%, and 22.05%, respectively; the average recall rates improved by 4.42%, 9.32%, 15.64%, and 24.92%, respectively; and the mean average precision improved by 3.11%, 4.15%, 8.02%, and 19.33%, respectively. The results demonstrate that the YOLOv8n-seg-CAA-BiFPN network can accurately achieve Chinese character stroke segmentation. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

14 pages, 3618 KiB  
Article
DBCW-YOLO: A Modified YOLOv5 for the Detection of Steel Surface Defects
by Jianfeng Han, Guoqing Cui, Zhiwei Li and Jingxuan Zhao
Appl. Sci. 2024, 14(11), 4594; https://doi.org/10.3390/app14114594 - 27 May 2024
Viewed by 561
Abstract
In steel production, defect detection is crucial for preventing safety risks, and improving the accuracy of steel defect detection in industrial environments remains challenging due to the variable types of defects, cluttered backgrounds, low contrast, and noise interference. Therefore, this paper introduces a [...] Read more.
In steel production, defect detection is crucial for preventing safety risks, and improving the accuracy of steel defect detection in industrial environments remains challenging due to the variable types of defects, cluttered backgrounds, low contrast, and noise interference. Therefore, this paper introduces a steel surface defect detection model, DBCW-YOLO, based on YOLOv5. Firstly, a new feature fusion strategy is proposed to optimize the feature map fusion pair model using the BiFPN method to fuse information at multiple scales, and CARAFE up-sampling is introduced to expand the sensory field of the network and make more effective use of the surrounding information. Secondly, the WIoU uses a dynamic non-monotonic focusing mechanism introduced in the loss function part to optimize the loss function and solve the problem of accuracy degradation due to sample inhomogeneity. This approach improves the learning ability of small target steel defects and accelerates network convergence. Finally, we use the dynamic heads in the network prediction phase. This improves the scale-aware, spatial-aware, and task-aware performance of the algorithm. Experimental results on the NEU-DET dataset show that the average detection accuracy is 81.1, which is about (YOLOv5) 6% higher than the original model and satisfies real-time detection. Therefore, DBCW-YOLO has good overall performance in the steel surface defect detection task. Full article
Show Figures

Figure 1

Back to TopTop