Evaluation of Power Insulator Detection Efficiency with the Use of Limited Training Dataset
Abstract
:1. Introduction
2. Recording of Visual Data
- The need to use advanced methods of image processing (deep learning);
- The need to build appropriate datasets which enable the construction of models describing the essential elements of technical infrastructure;
- A high degree of complication in making decisions concerning the assessment of the state of the examined object.
3. Application of the Proposed Approach to Insulator Detection
3.1. Dataset Preparation for CNN
- A large number of samples;
- Balance in terms of individual classes;
- High quality of images;
- Diversity in terms of depicting the analyzed objects (position, rotation, scale, lighting, obscuration);
- Accuracy of annotations (objects marking) in images.
- The diversity of data sources brings abundant data types and complex data structures, which increase the difficulty of data integration;
- Data volume is tremendous, and it is difficult to evaluate data quality within a reasonable amount of time;
- Data change very fast, and the “timeliness” of data is very short, which necessitates higher requirements for processing technology;
- No unified and approved data quality standards exist, and the research on the quality of large datasets only just began.
3.2. Description of the Proposed Approach
3.3. Training and Test Dataset
3.4. Applied Convolutional Neural Networks
3.5. The Analysis Results for All the Applied Convolutional Neural Networks
3.6. The In-Depth Analysis Results for the Faster R-CNN and R-FCN Networks
4. Conclusions
5. Discussion
- Flat and limited (unidirectional) views of objects. For aerial photographs taken perpendicular to the ground, objects of interest are relatively small and have fewer elements, mainly in flat and rectangular form. They usually include shots from above, omitting many important features of those objects in other planes.
- Large sizes of digital images. Currently, aerial imaging provides visual material at very high resolutions, which allows capturing increasingly more details, but at the same time introduces problems related to the use of sufficient computing power necessary to process them. These problems are eliminated by applying various methods of pre-processing for aerial photography. However, their proper selection requires a lot of research to determine the effectiveness of various solutions dedicated to specific technical problems.
- Overlapping objects. Objects may be occluded by other objects of the same type, which causes, e.g., inaccuracies when labeling data.
- Replication of the same objects in different digital images. The same object can occur in two separate images, which can lead to double detection and errors when recognizing objects in images.
Author Contributions
Funding
Conflicts of Interest
References
- Michalski, P.; Ruszczak, B.; Tomaszewski, M. Convolutional Neural Networks Implementations for Computer Vision. In Biomedical Engineering and Neuroscience, Advances in Intelligent Systems and Computing Book Series AISC; Springer: Cham, Switzerland, 2018; Volume 720, pp. 98–110. [Google Scholar] [CrossRef]
- Michalski, P.; Ruszczak, B.; Lorente, P.J.N. The Implementation of a Convolutional Neural Network for the Detection of the Transmission Towers Using Satellite Imagery. In Information Systems Architecture and Technology: Proceedings of 40th Anniversary International Conference on Information Systems Architecture and Technology–ISAT 2019; Springer International Publishing: Cham, Switzerland, 2020. [Google Scholar] [CrossRef]
- Pierdicca, R.; Malinverni, E.S.; Piccinini, F.; Paolanti, M.; Felicetti, A.; Zingaretti, P. Deep convolutional neural network for automatic detection of damaged photovoltaic cells. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 42, 893–900. [Google Scholar] [CrossRef] [Green Version]
- Zhang, S.; Zhang, S.; Wang, B.; Habetler, T.G. Deep Learning Algorithms for Bearing Fault Diagnostics—A Comprehensive Review. IEEE Access 2019, 8, 29857–29881. [Google Scholar] [CrossRef]
- Gasz, R.; Ruszczak, B.; Tomaszewski, M.; Zator, S. The Registration of Digital Images for the Truss Towers Diagnostics. In Information Systems Architecture and Technology: Proceedings of 39th International Conference on Information Systems Architecture and Technology–ISAT 2018; Advances in Intelligent Systems and Computing, (AISC, Volume 852); Borzemski, L., Świątek, J., Wilimowska, Z., Eds.; Springer: Cham, Switzerland, 2019. [Google Scholar] [CrossRef]
- Tomaszewski, M.; Ruszczak, B.; Michalski, P.; Zator, S. Detection of Power Line Insulators on Digital Images with the Use of Laser Spots. IET Image Process. 2019, 13, 2358–2366. [Google Scholar] [CrossRef]
- Yan, K.; Huang, S.; Song, Y.; Liu, W.; Fan, N. Insulator recognition based on convolution neural network. In Proceedings of the Chinese Control Conference CCC, Dalian, China, 26–28 July 2017; pp. 4077–4081. [Google Scholar]
- Zhao, Z.; Zhen, Z.; Zhang, L.; Qi, Y.; Kong, Y.; Zhang, K. Insulator Detection Method in Inspection Image Based on Improved Faster R-CNN. Energies 2019, 12, 1204. [Google Scholar] [CrossRef] [Green Version]
- Tomaszewski, M.; Osuchowski, J.; Debita, Ł. Effect of Spatial Filtering on Object Detection with the SUF Algorithm. In Proceedings of the International Scientific Conference BCI 2018 Opole, Opole, Poland, 13–14 March 2018; Springer: Cham, Swizerlands, 2018; pp. 121–140. [Google Scholar]
- Siddiqui, Z.A.; Park, U.; Lee, S.W.; Jung, N.J.; Choi, M.; Lim, C.; Seo, J.H. Robust Powerline Equipment Inspection System Based on a Convolutional Neural Network. Sensors (Basel) 2018, 18, 3837. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Shakhatreh, H.; Sawalmeh, A.; Al-Fuqaha, A.; Dou, Z.; Almaita, E.; Khalil, I.; Othman, S.; Khreishah, A.; Guizani, M. Unmanned Aerial Vehicles: A Survey on CivilApplications and Key Research Challenges. Available online: https://arxiv.org/abs/1805.00881 (accessed on 2 February 2020).
- Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef] [Green Version]
- Wang, Z.; Oates, T. Encoding time series as images for visual inspection and classification using tiled convolutional neural networks. In Proceedings of the Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2015, arXiv:1409.1556. arXiv.org e-Print archive. Available online: https://arxiv.org/abs/1409.1556 (accessed on 3 February 2020).
- Everingham, M.; Eslami, S.A.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes challenge: A retrospective. Int. J. Confl. Violence (IJCV) 2015, 111, 98–136. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Cai, L.; Zhu, Y. The Challenges of Data Quality and Data Quality Assessment in the Big Data Era. Data Sci. J. 2015, 14, 2. [Google Scholar] [CrossRef]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the IEEE Transactions on Pattern Analysis and Machine Intelligence; IEEE: Piscataway, NJ, USA, 2016; Volume 39, pp. 1137–1149. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition; IEEE: Piscataway, NJ, USA, 2014; pp. 580–587. [Google Scholar]
- Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; IEEE: Piscataway, NJ, USA, 2015; Volume 2015, pp. 1440–1448. [Google Scholar]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A.C. SSD: Single Shot Multibox Detector, Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). In Proceedings of the Computer Vision – ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 4510–4520. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement, 2018. arXiv 2018, arXiv:1804.02767. Available online: https://arxiv.org/abs/1804.02767 (accessed on 3 February 2020).
- Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object Detection via Region-based Fully Convolutional Networks. In Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
Name of the CNN | Abbreviation | CNN Model | Main Features |
---|---|---|---|
Faster region convolutional neural network | Faster R-CNN |
|
|
You Only Look Once v3 | YOLO v3 |
| |
Single-Shot MultiBox Detector Inception | SSD Inception |
| |
Single Shot MultiBox Detector Lite MobileNet v2 | SSD MobileNet |
|
|
Region-based fully convolutional network | R-FCN |
|
|
Network Model | Number of Frames | Average Precision IOU = 0.25 | Precision IOU = 0.25 | Recall IOU = 0.25 | Average Precision IOU = 0.5 | Precision IOU = 0.5 | Recall IOU = 0.5 | Average Precision IOU = 0.75 | Precision IOU = 0.75 | Recall IOU = 0.75 |
---|---|---|---|---|---|---|---|---|---|---|
Faster R-CNN | 15 | 0.6156 | 0.8846 | 0.6137 | 0.3952 | 0.6537 | 0.4535 | 0.1064 | 0.1509 | 0.1047 |
Faster R-CNN | 30 | 0.7044 | 0.8989 | 0.7223 | 0.4723 | 0.7000 | 0.5625 | 0.0670 | 0.1731 | 0.1391 |
Faster R-CNN | 60 | 0.8043 | 0.9274 | 0.8328 | 0.6718 | 0.7964 | 0.7152 | 0.1584 | 0.2671 | 0.2398 |
R-FCN | 15 | 0.7069 | 0.9123 | 0.7027 | 0.4805 | 0.7110 | 0.5477 | 0.1110 | 0.1668 | 0.1285 |
R-FCN | 30 | 0.8015 | 0.9272 | 0.8461 | 0.5580 | 0.7354 | 0.6711 | 0.1178 | 0.2098 | 0.1914 |
R-FCN | 60 | 0.8948 | 0.9515 | 0.9266 | 0.6598 | 0.8139 | 0.7926 | 0.1589 | 0.2720 | 0.2648 |
YOLO v3 | 15 | 0.2651 | 0.9324 | 0.2965 | 0.1418 | 0.5258 | 0.1672 | 0.0056 | 0.0590 | 0.0188 |
YOLO v3 | 30 | 0.3493 | 0.9208 | 0.3727 | 0.1127 | 0.4797 | 0.1941 | 0.0114 | 0.0531 | 0.0215 |
YOLO v3 | 60 | 0.4445 | 0.9427 | 0.4500 | 0.2000 | 0.5483 | 0.2617 | 0.0909 | 0.0687 | 0.0328 |
SSD Inception | 15 | 0.3589 | 0.9479 | 0.3910 | 0.2285 | 0.6771 | 0.2793 | 0.0909 | 0.1752 | 0.0723 |
SSD Inception | 30 | 0.4486 | 0.9528 | 0.4813 | 0.3096 | 0.7301 | 0.3688 | 0.0331 | 0.1570 | 0.0793 |
SSD Inception | 60 | 0.5365 | 0.9589 | 0.5465 | 0.3944 | 0.7615 | 0.4340 | 0.1199 | 0.2557 | 0.1457 |
SSD MobileNet | 15 | 0.3520 | 0.9286 | 0.3555 | 0.2238 | 0.6224 | 0.2383 | 0.0273 | 0.1327 | 0.0508 |
SSD MobileNet | 30 | 0.4344 | 0.9211 | 0.4238 | 0.2956 | 0.6842 | 0.3148 | 0.0303 | 0.1171 | 0.0539 |
SSD MobileNet | 60 | 0.5290 | 0.9225 | 0.5207 | 0.3014 | 0.6657 | 0.3758 | 0.0268 | 0.1488 | 0.0840 |
Network Model | Number of Frames | Average Precision IOU = 0.5 | Precision IOU = 0.5 | Recall IOU = 0.5 | Average Precision IOU = 0.5, Expanded Training | Precision IOU = 0.5, Expanded Training | Recall IOU = 0.5, Expanded Training |
---|---|---|---|---|---|---|---|
Faster R-CNN | 15 | 0.4712 | 0.7251 | 0.5379 | 0.4384 | 0.8933 | 0.4676 |
Faster R-CNN | 30 | 0.5714 | 0.7574 | 0.6574 | 0.6106 | 0.8774 | 0.6039 |
Faster R-CNN | 60 | 0.6672 | 0.7929 | 0.7105 | 0.6221 | 0.9061 | 0.6445 |
R-FCN | 15 | 0.4060 | 0.7424 | 0.4863 | 0.4368 | 0.9015 | 0.4363 |
R-FCN | 30 | 0.4852 | 0.7650 | 0.5813 | 0.4402 | 0.9204 | 0.4969 |
R-FCN | 60 | 0.6609 | 0.8206 | 0.7738 | 0.6109 | 0.9068 | 0.6766 |
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Tomaszewski, M.; Michalski, P.; Osuchowski, J. Evaluation of Power Insulator Detection Efficiency with the Use of Limited Training Dataset. Appl. Sci. 2020, 10, 2104. https://doi.org/10.3390/app10062104
Tomaszewski M, Michalski P, Osuchowski J. Evaluation of Power Insulator Detection Efficiency with the Use of Limited Training Dataset. Applied Sciences. 2020; 10(6):2104. https://doi.org/10.3390/app10062104
Chicago/Turabian StyleTomaszewski, Michał, Paweł Michalski, and Jakub Osuchowski. 2020. "Evaluation of Power Insulator Detection Efficiency with the Use of Limited Training Dataset" Applied Sciences 10, no. 6: 2104. https://doi.org/10.3390/app10062104