Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Investigating the Performance of Gesture-Based Input for Mid-Air Text Entry in a Virtual Environment: A Comparison of Hand-Up versus Hand-Down Postures
Previous Article in Journal
Thermal Ablation and High-Resolution Imaging Using a Back-to-Back (BTB) Dual-Mode Ultrasonic Transducer: In Vivo Results
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Automatic Concrete Crack-Detection Method Fusing Point Clouds and Images Based on Improved Otsu’s Algorithm

1
School of Water Conservancy Science & Engineering, Zhengzhou University, Zhengzhou 450001, China
2
School of Geo-Science & Technology, Zhengzhou University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(5), 1581; https://doi.org/10.3390/s21051581
Submission received: 30 January 2021 / Revised: 15 February 2021 / Accepted: 20 February 2021 / Published: 24 February 2021
(This article belongs to the Section Remote Sensors)

Abstract

:
Cracks are one of the main distresses that occur on concrete surfaces. Traditional methods for detecting cracks based on two-dimensional (2D) images can be hampered by stains, shadows, and other artifacts, while various three-dimensional (3D) crack-detection techniques, using point clouds, are less affected in this regard but are limited by the measurement accuracy of the 3D laser scanner. In this study, we propose an automatic crack-detection method that fuses 3D point clouds and 2D images based on an improved Otsu algorithm, which consists of the following four major procedures. First, a high-precision registration of a depth image projected from 3D point clouds and 2D images is performed. Second, pixel-level image fusion is performed, which fuses the depth and gray information. Third, a rough crack image is obtained from the fusion image using the improved Otsu method. Finally, the connected domain labeling and morphological methods are used to finely extract the cracks. Experimentally, the proposed method was tested at multiple scales and with various types of concrete crack. The results demonstrate that the proposed method can achieve an average precision of 89.0%, recall of 84.8%, and F1 score of 86.7%, performing significantly better than the single image (average F1 score of 67.6%) and single point cloud (average F1 score of 76.0%) methods. Accordingly, the proposed method has high detection accuracy and universality, indicating its wide potential application as an automatic method for concrete-crack detection.

1. Introduction

Cracks are one of the main types of distress in concrete surfaces, with a significant impact on the bearing capacity and durability of concrete structures. Therefore, an automatic and accurate method for recognizing them is crucial for the monitoring and maintenance of concrete structures. In the past few years, various technologies have been developed for automatic concrete distress detection. According to the principles of detection, we can divide these into three main categories: 2D image-based detections, 3D laser-scanning methods, and 2D–3D combination techniques.
Many researchers have proposed various crack-edge detection algorithms based on 2D images, such as Laplacian, Sobel, Canny, and Prewitt [1,2,3,4]. These are effective in detecting cracks with obvious edge-gradient changes but are more sensitive to noise. Otsu et al. [5] proposed an automatic threshold image segmentation method based on the maximum variance (in gray) between classes. This method could calculate an appropriate threshold automatically but performs poorly in the presence of shadows or stains. Nguyen et al. [6] comprehensively considered features of cracks such as the grayscale values, texture, profile shape, and other features, achieving more accurate crack detection. A method based on deep learning has also been used for image crack detection [7,8,9,10,11,12,13], but this method requires many training samples. Although the aforementioned 2D image-based techniques provide various methods for automatic crack detection in specific cases, they are unable to distinguish cracks from artifacts such as illumination and water or oil stains, which is a major problem.
With the recent developments and innovations in hardware, 3D-laser disease-detection technology has become a research hotspot [14,15,16,17,18]. By contrast with the principle of 2D imaging, 3D-laser technology measures the geometric parameters of a target object’s surface and generates abundant high-precision spatial information called a point cloud [19,20], so crack detection is not interfered with by the above factors. There are two main methods for extracting cracks from 3D point clouds. (1) projecting dense 3D point clouds into 2D images and extracting cracks using image-processing techniques [21,22]; and (2) extracting cracks directly from the 3D point clouds [23,24]. Based on the first method, Yang et al. [25] used the reflectance information from point clouds to detect cracks in a tunnel structure through morphology and a Canny edge detection operator, but there were many fractures in the detected cracks. Guan et al. [26] proposed an iterative tensor voting method for the high-precision detection of road cracks. However, this approach was mainly developed to extract the types and locations of pavement cracks rather than the crack widths. Jiang et al. [27] used the inverse distance weighting method to generate point cloud raster images to realize crack detection for post-earthquake buildings. However, this required many shape features to remove false cracks. Based on the second method (direct extraction). Yu et al. [28] used intensity information from the point clouds to directly extract the crack skeleton. Firstly, Otsu’s method was used to extract candidate cracks and, then, a spatial density filter was used to denoise them. Finally, crack skeletons were extracted based on an L1-medial skeleton extraction method. This avoided a time-consuming neighborhood search in dense discrete point clouds and was fast to execute. However, crack points account for only a small proportion of mobile laser scanning data, which made Otsu’s threshold unreliable for segmented cracks. According to the scanning angle and time of the scanner, Zhong et al. [29] established a 2D index for each point cloud to reduce the dimensionality in a non-destructive manner, and reflectance and depth information were then used to extract cracks. Since the density of a 3D point cloud is not high enough, existing 3D methods often make it difficult to detect small cracks.
Considering the complex conditions of concrete surfaces, such as the presence of stains and shadows, it is difficult to achieve high accuracy using a 2D-image method or a 3D point clouds separately. Therefore, some scholars have proposed methods combining 2D and 3D crack extraction [30]. Valença et al. [31] used coordinate information from the point clouds to orthorectify images, solving the problems encountered when applying image-detection technology to detecting cracks in large buildings. However, the method was still based on the image for crack detection and did not make full use of the depth information of point clouds. Based on the Dempster–Shafer (D–S) theory, Huang et al. [32] proposed a new road-crack detection method combining 2D images and 3D-laser-scanning data. Their experimental results showed that this method could improve the accuracy of crack detection and reduce the fallout ratio. However, it only used four structured lights in the same direction to judge whether a crack in the image was real, increasing the possibility of miscalculation.
To overcome the discussed limitations for high precision concrete-surface crack detection based on single 2D images and 3D point clouds methods, an automatic method fusing 3D point clouds and 2D images based on an improved Otsu’s method is proposed in this paper. In this method, depth information from point clouds and gray information is fused at the pixel level; then based on an improved Otsu segmentation method and fine detection of cracks, high-precision crack detection, is achieved. The main innovations of the proposed method in this paper are as follows: (1) a new fusion method of 3D point clouds and 2D image data is proposed; (2) proposing an improved Otsu crack detection method; (3) discussing the application of two different types of 3D laser scanners in the field of crack detection. The rest of the article is organized as follows: the proposed fusion detection method is explained in detail in Section 2, experiments results and discussion are shown in Section 3, and conclusions are given in Section 4.

2. Methods

In this study, the overall framework of the proposed method is mainly composed of the following steps: (1) data preprocessing; (2) the fusion of 3D point clouds and 2D images; (3) crack detection based on an improved Otsu’s method; (4) denoising and repair of cracks. Among them, the fusion of 3D point clouds and 2D images and the crack detection based on the improved Otsu’s method are the two most important parts. The former determines the quality of the fusion image and is the basis for high-precision crack detection, while the latter determines the accuracy with which cracks are extracted. A flow chart summarizing the process is presented in Figure 1.

2.1. Data Preprocessing

Due to interference from the sensor itself, the environment, and human factors, there is a certain amount of noise in the collected point cloud and image data, so they need to be preprocessed separately. For the images, the three-color component weighted average method is first applied for image graying: the true color image (red, green and blue, RGB) is converted into grayscale by removing a lot of color information that is not good for crack detection, which reduces the image data. Specifically, the 24-bit RGB data ((28)3 = 16,777,216 colors) is converted into 8-bit gray-level (28 = 256 levels) data. Image noise is then removed using the median filter to avoid its interference in crack detection, (the median filter was selected after comparing various filtering methods). For the 3D point clouds, the preprocessing mainly includes two steps: (1) the filtering and denoising of the point clouds; and (2) depth-image acquisition. Before fusing the 2D and 3D data, the point clouds need to be reduced into 2D depth images. However, any noise in the original point clouds greatly impacts the quality of the generated depth images; therefore target-area point clouds are extracted from the original point clouds and then denoised with the median filter.
To obtain the depth image, it is necessary to transform the filtered point clouds from the original coordinate system to the new coordinate system with the concrete surface fitting plane as the XOY plane. Firstly, principal component analysis (PCA) is used to fit the point clouds and calculate the parameters of the fitting plane. Next, the rotation and translation matrix of the fitting plane transformed into the XOY plane is calculated. Then, the coordinate of all the point clouds is transformed using the calculated rotation-translation matrix. The points from plane fitting contain crack points, which cause errors in the plane fitting, so the Z coordinate threshold was used to remove the crack point clouds in this study; then, a second plane fitting for the remaining non-crack point clouds was carried out by the PCA method, and the corresponding rotation translation matrix was calculated again. Finally, the rotation and translation matrix was used to transform the coordinates of all the point clouds including the cracked and non-cracked point clouds.
The depth of a crack point can be directly expressed by the Z coordinate value of point clouds in the new coordinate system, so point clouds can be projected into depth images as follows:
(1)
An image grid of size m × n , as shown in Figure 2 is established, and the edge length of the image pixel is P s (the grid spacing). The relationship between the image size and edge length of the pixel is as follows:
m = y max y min P s ,
n = x max x min P s ,
where y max is the maximum Y coordinate; y min is the minimum Y coordinate; x max is the maximum X coordinate, and x min is the minimum X coordinate. P s refer to the average point distance of point clouds.
(2)
The number of points falling into the grid is calculated, and the average of their coordinate values is taken as the depth.
z ¯ i , j = i n z i , j n ,
where z i , j is the depth value of row and column of the image grid; z ¯ i , j is the number of points falling into the grid and the average depth value of row and column of the image grid. The reason why the average value is chosen instead of the maximum or minimum value is to avoid error.
(3)
The depth is normalized to a 0–255 scale and then the image grid matrix is output as the depth image.
G i , j = z ¯ i , j z ¯ min z ¯ max z ¯ min × 255 ,
where z ¯ i , j is the average depth value of row and column of the image grid; z ¯ max is the maximum average depth value in the image; z ¯ min is the minimum average depth value in the image.

2.2. Fusion of 3D Point Clouds and 2D Gray Images

The fusion of 3D point clouds and 2D images is one of the key aspects of our method. The purpose is to generate a fused image with low noise and prominent cracks by fusing the depth of the point clouds and the image gray information. In this study, the point clouds were projected into depth images, and then, the latter and the 2D gray images were fused by an image processing method mainly consisting of: (1) the registration of the depth and gray image; (2) the fusion of the depth and gray image.

2.2.1. Registration of Depth and Gray Image

Image registration means overlaying two or more images of the same scene taken at different times, from different viewpoints, and/or by different sensors. It geometrically aligns two images—moving and fixed images. The key is to find the best transformation matrix for maximizing the degree of alignment between the two images, and it mainly involves the calculation of the control points coordinates and geometric transformation. The advantage of the method in this study is that it can not only achieve accurate image registration but also use the target to determine the ground sampling distance (GSD) in the images (the GSD means the actual distance corresponding to the pixel).
Firstly, we selected circular targets uniformly distributed on the surface of the concrete structure as the control points and calculated their image coordinates The Hough transform circle detection algorithm [33] was used to obtain the center coordinates of the circular target in the 2D images. As for the point clouds, because the 3D laser scanner can directly obtain the precise 3D coordinates of the target center, the corresponding image coordinates can be obtained by calculating the location of the target center point in the depth image.
After obtaining the precise control point coordinates of the two images, the precise registration of the depth image and the 2D image can be realized through geometric transformation between the two images, which includes spatial geometric transformation and grayscale interpolation. The former defines the pixel distribution of the image in the new coordinate system; the latter assigns the pixel gray level of the image after spatial transformation (grayscale interpolation). This study used a bilinear interpolation method whose principle is shown in Figure 3.
I 2 = g I 1 f x , y ,
where I 1 is the reference image; I 2 is the image to be registered; f represents the geometric transformation of the 2D space; g represents the one-dimensional grayscale transformation.
To following measures were taken to ensure the accuracy of the registration: (1) a certain number of targets were evenly distributed on the surface of concrete; (2) the center coordinates of the target were obtained by using the circular detection algorithm and fine scan; (3) the two images after registration were overlapped and displayed to ensure that the crack edges completely overlapped. The registration results for the depth and gray image are shown in Figure 4; the depth image is displayed in color in Figure 4b to clarify the registration effect. Figure 4c is the registration effect picture of the two images superimposed. It can be seen from Figure 4c that the crack areas of the two images overlapped well, and the registration result was good.

2.2.2. Fusion of Depth and Gray Image

The fusion of depth and gray images is based on their complementarity for crack detection. For example, in a background area with stain interference, the grayscale of the gray image is smaller than the normal grayscale value, but the grayscale of the depth image is not affected; in an area with muddy landfill or shallow cracks, the grayscale of the depth image is larger than the normal grayscale value, but the grayscale of the gray image is not affected. Harnessing the advantage of the two types of data, and according to the concrete surface background interference and the depth and width of the cracks, we used the weighted-average method of pixel-level fusion to assign different weights to the depth and gray image for data fusion. The fusion image generated by the above steps is a composite of gray information and depth information. Compared with a single image or depth image, it can have reduced background noise and enhanced crack information, providing a solid foundation for subsequent crack detection research.
F ( i , j ) = ω 1 A ( i , j ) + ω 2 B ( i , j ) ,
where A represents the depth image; B represents the image, and ( i , j ) represents the coordinates of pixels in the image; ω 1 , ω 2 is the weighting factor and usually take ω 1 + ω 2 = 1 .
The comparison of the different images is shown in Figure 5. It can be seen by comparing Figure 5a,c, that the noise from the original image is reduced, and the crack information is enhanced. By comparing Figure 5b,c, it can be observed that the crack information in the fusion image is enhanced, as expected.

2.3. Rough Detection of Cracks Based on Improved Otsu’s Method

Otsu’s method is an adaptive threshold crack-detection method that is widely employed. The basic principle is dividing the image into background and target according to a threshold. The best segmentation threshold is the pixel value that maximizes the variance between the two kinds of pixel. Although a fusion image has more advantages for crack detection than a single image or depth image, with the traditional Otsu method, it still presents certain limitations in complex backgrounds and images with unclear background contrasts.
To overcome the shortcomings of the traditional Otsu method in this regard, this study proposes an improved Otsu crack-detection method that incorporates several constraints for the image-segmentation algorithm, which makes the block-based crack segmentation more accurate. This method first divides the fused image into blocks and calculates the segmentation thresholds for different image sub-blocks. Next, several constraints are introduced to divide the image sub-blocks into the crack and background sub-blocks, and then, the crack sub-blocks are roughly extracted through the steps described above; lastly, all the binary sub-block images are merged, and the fine detection of cracks is realized by denoising and repairing cracks. A flow chart summarizing the improved Otsu crack detection method is shown in Figure 6a, and the specific steps are as follows:
Firstly, the fused image is divided into blocks, and then, the Otsu’s threshold for all the image sub-blocks are calculated. Compared with a single threshold for the whole image, different thresholds for different image sub-blocks can allow the extraction of the cracks from the background with more accuracy.
However, the forced binarization to the background sub-blocks would result in errors. Therefore, we introduced local constraints related to the image sub-blocks such as the pixel ratio P r , c , local gray average μ r , c , and the difference in gray average d r , c , and global constraints related to the entire image, such as the global gray average μ and global gray standard deviation σ . Based on these constraints, it is judged whether the image sub-block contains cracks. If so, the Otsu threshold segmentation is performed on the image sub-block; otherwise, the grayscale values of all the pixels of the image sub-blocks are set to 0 (background sub-block: no crack). The reason these constraints are introduced is that the proportion of crack pixels in the whole image is small, the difference between the mean gray value of the crack pixels and that of the background pixels is large, and the mean gray value of the image sub-blocks generally fluctuates around the mean gray value of the whole image. In short, the above constraints help distinguish the cracks and background in the image. The specific judgment process for image sub-blocks is shown in Figure 6b, and the specific definitions of the constraints are as follows:
P r , c = n r , c N r , c ,
where r , c is the number of the image sub-block; n r , c is the number of pixels cracked under the threshold t r , c in the image sub-block; N r , c is the number of all pixels in the image sub-block.
d r , c = μ 1 r , c μ 2 r , c ,
where μ 1 r , c is the gray average of the background of the image sub-block under the threshold t r , c ; μ 2 r , c is the gray average of the crack of the image sub-block under the threshold t r , c .
Figure 7 is a comparison between the traditional and improved Otsu methods for crack based on fusion images. Comparing Figure 7b,c, it can be seen that the traditional Otsu method recognizes many background interferences as cracks, resulting in many errors. Meanwhile, the improved Otsu method proposed in this paper can effectively avoid interference from complex backgrounds and accurately detect cracks.

2.4. Denoising and Connection of Cracks

After the rough crack detection, some crack burrs, noises, and fractures may still exist, which could seriously affect the accuracy. Hence, denoising and connection of cracks are needed to refine the final detection. The cracks in the fusion image mainly present the following two characteristics: (1) generally linear shape; (2) areas are large relative to the noises. Therefore, the connected component labeling method can be used to remove noises based on the shape and area feature constraint of the connected component. Then the morphology method is used to remove the crack burrs and connect the broken cracks. The specific process is shown in Figure 8, and the specific steps are as follows:
Firstly, connected region analysis is performed on the binary image after the implementation of the improved Otsu’s method. First, employing the algorithm used in the connected-region labeling function “bwlabel” in MATLAB, the pixels that conform to the four-connection relation are classified into the same connected domain. For extracting the crack, the connected area and minimum length /width ratio of the circumscribed rectangle of the crack should meet the conditions of Equations (9) and (10).
A i A t i ,
where A i is the area of the connected region; A t i is the area threshold of the connected region conforming to the crack condition.
K i = L i W i R K ,
where K i is the length to width ratio of the connected domain; L i is the length of the connected domain; W i is the width of the connected domain, and R K is the threshold of the length-width ratio of the connected domain satisfying the condition of being a crack.
After the noise is removed based on the above constraints, the open and close operations for morphology can be used to remove the burr and connect the fractured cracks. The crack burrs and isolated noise are removed using the open operation, and the holes in the cracks are filled and the cracks in adjacent areas are connected using the closed operation. Figure 9 is a schematic diagram of the effects before and after the denoising and connection of the crack.
A comparison of before and after the denoising and connection of cracks is shown in Figure 10, where it can be observed that the noise and burrs of the crack are removed.

3. Experimental Results and Analysis

The concrete crack data in this experiment were collected on different pavements and walls at a university covering an area of 3 square kilometers. The cracks include 3D point clouds and 2D-image data. The 3D point clouds were collected by a handscan700 handheld 3D laser scanner and RIEGL VZ-2000 terrestrial 3D laser scanner. The two scanners have different spatial resolutions: the minimum for the handheld scanner is 0.2 mm (the minimum distance between two points), while the point cloud of the terrestrial laser scanner is related to the scanning distance and angular resolution. The shorter the distance, the smaller the angular resolution and the higher the accuracy. To obtain as much 3D information of the crack as possible, we used the highest accuracy that the scanner can achieve, and its corresponding minimum point cloud resolution was 1 mm. The specific technical parameters of the scanner are shown in Table 1 and Table 2.
The 2D images were captured by a camera with a resolution of 7952 × 5304, orthogonally to the concrete surface. The 3D laser scanner and camera were completely separated, each acquiring data separately, so that high-resolution concrete-crack images could be obtained at a close distance.
Three typical cracks were included: transverse, longitudinal, and inclined. The length and width of the crack concrete area measured by the hand-held 3D laser scanner are about 0.5 m, and the grid spacing of the depth image of the point clouds collected with the handheld 3D laser scanner is 0.5 mm. Therefore, the length and width of the corresponding depth image are about 500 × 500. The length and width of the crack concrete area measured by the terrestrial 3D laser scanner are about 0.6–1.5 m, and the grid spacing of the depth image of the point clouds collected by the terrestrial 3D laser scanner is 1 mm in the experiment. Thus, the length and width of the corresponding depth image were about 600–1500. The relevant parameter settings used for this method are shown in Table 3.
To verify the feasibility and universality of the proposed method, it was compared with a single 3D point clouds method and a single 2D-images method, respectively. All the real cracks were obtained manually for evaluation. Some of the crack detection results are as follows: Figure 11 shows the results for a group of longitudinal cracks with detection by three methods, where the point clouds data came from a handheld laser scanner. Figure 12 shows the results of a group of inclined cracks by three methods with detection, where the point clouds data came from a terrestrial laser scanner. Figure 11d and Figure 12d are the 2D image-detection results; upon comparison with the manual annotation results (Figure 11c and Figure 12c), much false detection was observed, due to there being many stains near the cracks Figure 11e and Figure 12e show the results of detection based on 3D point clouds. Although 3D point clouds are not affected by crack surface stains, the detection of small or shallow cracks is poor, resulting in fractures. Figure 11f and Figure 12f show the results of detection using the method proposed in this paper. Those shown in Figure 11f are largely consistent with the manual annotations, with no fractures in the cracks. This demonstrates that the proposed method can overcome interference from the background, with a significantly reduced rate of false detection, while still small cracks. Although the background interference was also removed as shown in Figure 12f, some cracks were not detected due to the influence of various factors, such as cracks not being obvious because of landfills, the grayscale values of stains, and background being very similar, and cracks being too small. Moreover, the point clouds of the crack in Figure 12 come from the terrestrial 3D laser scanner which is less accurate than the handheld laser scanner, so the ability to recognize the small cracks shown in Figure 12f is not as good as Figure 11f.
To illustrate the superiority of the proposed method, a set of cracks was randomly selected from large experimental samples, and the deep learning crack detection algorithm called DeepCrack [34] and the method in this paper were compared. The detection results are shown in Figure 13. It can be seen from Figure 13a that there is a large area of background interference near the crack. Comparing Figure 13b,c, we can see that the method in this paper can resist noise interference and accurately extract cracks. The deep learning method recognizes large crack edge pixels as background pixels, resulting in the width of the crack detected being much smaller than the true width.
In this study, the precision, recall, and F1-measure (also known as F1 score) were used to quantitatively analyze and evaluate the performance of the proposed method. The precision measures the exactness or fidelity of detection, while the recall describes the completeness of detection. We used a common evaluation standard for the proposed method, the F1-measure (also known as F1 score), which combines precision and recall. The closer the F1 value is to 1, the better the method performs. P, R, and F1 were calculated using Equations (11)–(13), respectively. Table 3 shows the quantitative results.
P = T P T P + F P ,
R = T P T P + F N ,
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l ,
where T P denotes true positives; that is, pixels labeled as crack pixels are correctly recognized as crack pixels; F P denotes false positives; that is, pixels labeled as non-crack pixels are incorrectly recognized as crack pixels; F N represents false negatives; that is, pixels labeled as crack pixels are incorrectly detected as non-crack pixels.
Table 4 and Table 5 show the results of detection using three methods for three types of crack. The point cloud data for the cracks shown in Table 4 were obtained by a handheld laser scanner, while the data in Table 5 were obtained by a terrestrial laser scanner. Compared with the detection method using a single 2D image, as observable from the tables, the proposed crack detection method fusing point clouds and 2D images is significantly more precise. The biggest increase in precision is seen for Crack No.3 (Table 4), by 51.8% from 38.7% to 90.5%; the recall of the single 3D point-cloud detection method fluctuates between 46.8% and 88.5% (the results for five of the six groups of data are below 80%), while that of the proposed method is consistently above 80%. According to the comprehensive evaluation index, the F1 score, the proposed fusion method is better than the single 3D-point-cloud or 2D-image methods for five of the six groups of data containing three types of crack, with F1 scores consistently above 80%. Therefore, the proposed detection method shows great performance and universality.
Table 6 shows the average detection results for all the cracks; it can be seen that the proposed method achieved an average precision of 89.0%, and a recall of 84.8%; so it is highly accurate. The average F1 score was 86.7%, which is 19.1% and 10.7% higher than the scores of the single 2D-image and 3D point-clouds methods, respectively, showing that the proposed method shows the best overall performance.
The quantitative results detected by the deep learning and the proposed method are shown in Table 7. It can be seen from the table that both accuracy and recall rate and F1 score of the proposed method are superior to the DeepCrack method.
The crack-detection results for the handheld and the terrestrial 3D laser scanners are shown in Table 8. With the single 3D method, the three groups of cracks acquired by the handheld scanner are smaller than those acquired by the terrestrial laser scanner, but the resolution of the handheld scanner (0.5 mm) is higher than that of the terrestrial (1 mm). Therefore, the average precision (90.2%), recall (76.9%), and F1 score (82.9%) of the handheld laser scanner are better than those of the terrestrial. The results obtained from fusion with high-precision point clouds were expected to be better than those obtained from fusion with low-precision point clouds, but this was not the case. This could be because the background interference in the 2D image data of three groups corresponding to the handheld laser scanner is greater than that with the terrestrial laser scanner, compromising the detection results. Comparing the detection results from different types of point cloud shows that the accuracy of the point cloud has a greater impact on the precision of the posed fusion detection method and that higher-precision point clouds can lead to better detection results. Although the point clouds collected by the handheld 3D laser scanner are more accurate, the terrestrial 3D laser scanner is far superior in terms of scanning efficiency. Therefore, the advantages of both might be harnessed by choosing between them according to the scenario: for wide cracks or large scenes with low accuracy requirements, the terrestrial 3D laser scanner could be used to obtain the point cloud data, while in the small scenes with shallow cracks and cracks buried or with high precision requirements, the handheld 3D laser scanner could be used to obtain said data. In the future, with the continuous advancement of terrestrial 3D laser-scanning technology, more accurate and faster scanning will be realized, supporting the ability of the method in this paper to detect cracks with high-precision and high-efficiency.
Pixel calibration was realized by using a circular target. Then, the corresponding true distances (GSDs) of pixels for six groups of crack images were obtained, as shown in Table 9; it can be seen that the GSDs of the three groups of images corresponding to the handheld 3D laser scanner are better than those for the terrestrial 3D laser scanner. The GSD reflects the distance between the camera and the concrete surface. In theory, the closer the camera is to the concrete surface, the higher the image resolution and the more detailed the crack information obtained. It can be seen from Table 4, Table 5 and Table 6 that the GSDs of the images of Cracks No. 1–3 higher than those of the images of Cracks No. 4–6, but the F1 value is not optimal. We speculate that this is because there were too many stains on the surface of the concrete cracks, which were themselves accurately recorded in the images with high GSDs.
Table 10 describes the maximum width error between the results obtained with the method in this paper and the manual annotations; one can see that the wider crack, the lower the relative error. For cracks with a point cloud accuracy of 0.5 mm, when the crack width is about 2 cm, the relative error for the maximum width is within 7.3%; when the crack width is less than 1 cm, the relative error exceeds 10%; For cracks with a point cloud accuracy of 1 mm, when the crack width is about 2 cm, the relative error is within 5.6%; when the crack width is greater than 5 cm, the relative error is less than 2%.

4. Discussion

The weighted average pixel-level image fusion method is used to fuse the depth and gray image, so the selection of the factor weighting greatly impacts the final detection accuracy. If the depth image is overweighted, it becomes difficult to effectively take advantage of the high resolution of the image, which can allow the identification of tiny cracks. By contrast, if the weight of the image is too large, then stains and uneven illumination in the image can greatly interfere with crack detection. Therefore, appropriately setting these factors according to the situation of the concrete crack and the actual measured performance of the scanner will improve the accuracy of crack detection. The weighting factors in Table 3 are the best weighting factors we chose in the experiment. The crack comes from different concrete pavements and walls in the university campus, so the type, width, and concrete conditions of the cracks are very different. Therefore, the optimal weighting factors for different fractures are different. But we did give some guiding advice. For wide cracks with greater environmental interference, we recommend that the weighting factor Kr of the depth image should be greater than 0.45; for narrow cracks with greater environmental interference, it should be greater than 0.65. In future research, we will study the influence of weight factors in depth and find a weight factor allocation method with wider applicability.
The scanner distance, incident angle, and other factors have a certain impact on the accuracy of crack detection. The existing research of Kim [20] and the findings in this paper show that the farther the terrestrial 3D laser scanner is from the concrete surface, the greater the incident angle, the sparser the point cloud obtained, and the more difficult it is to detect points inside the crack. Therefore, we recommend measuring as close as possible and orthogonal to the concrete surface, but no closer than the closest distance of the scanner. For handheld 3D lasers, repeated measurements are performed within 25–30 cm of the diseased concrete surface to obtain all the point clouds. Therefore, it is worth considering performing repeated measurements from multiple angles to obtain as much depth information as possible inside the crack. In actual engineering applications, to reduce the time cost of data acquisition, we will consider using more efficient and convenient 3D laser sensors, such as the use of Unmanned Aerial Vehicle 3D laser scanners to collect point cloud if the point cloud meets the requirements of precision and density.
The algorithm used for crack extraction has a great influence on the accuracy of detection. The proposed method, fusing point clouds and images based on an improved Otsu algorithm, not only obtains width and depth information for cracks but is also strongly resistant to interference from stains. It should be mentioned that the development of computer vision technology has made deep learning in crack-detection research a hotspot in recent years. Therefore, in the future, we will conduct in-depth research on crack detection with the fusion of point clouds and images based on deep learning.

5. Conclusions

To deal with the disadvantages of low accuracy in extracting concrete surface cracks based on single 3D point-cloud and 2D-image data, this paper proposes an automatic concrete crack detection method fusing 3D point clouds and 2D images based on an improved Otsu’s algorithm. Firstly, depth images from point clouds and 2D images are fused at the pixel level to generate a fused image, and then accurate crack detection is realized based on an improved Otsu method. The main contributions of this paper to the research on crack detection, regarding the proposed method, are as follows: (1) the point clouds are first processed to reduce dimensionality not only does this significantly reduces the data and necessary calculations, but mature 2D image-processing technology can be used to detect cracks from the fusion image; (2) the fusion of 3D point clouds and 2D images is realized at the pixel level. The fusion image combines the grayscale information from the high-resolution images and the depth information from the point clouds, fully harnessing the advantages and compensating for the shortcomings of both; (3) an improved Otsu crack-detection method is proposed. By dividing the image into blocks and introducing constraints related to the image sub-blocks, the error that would be introduced by the forced binarization of the image sub-blocks is avoided, thereby improving the crack detection accuracy; (4) the effects on the detection accuracy of different methods for acquiring the point clouds were studied, and methods are proposed according to the scenarios and precision requirements for crack detection. Further research will seek to improve the image segmentation algorithm.

Author Contributions

Conceptualization, J.L. and X.C.; methodology, X.C.; software, X.C.; validation, J.L.; formal analysis, X.C.; investigation, X.C., S.H., P.L.; resources, J.L.; data curation, X.C.; writing—original draft preparation, X.C., Q.S.; writing—review and editing, X.C., J.L., Q.S.; visualization, X.C.; supervision, X.C.; funding acquisition, J.L. and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Foundation of He Nan Educational Committee under Item Number 15A420002 and Item Number 14A420002. This research was also funded by China Postdoctoral Science Foundation, grant number 2019M662534; and National natural science foundation, grant number 42001405.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Acknowledgments

Thanks are due to Shuowen Huang, Peirong Liu for his data collection and technological help, and to Yao Cao for assistance with the experiments. We are grateful to anonymous reviewers for their thoughtful comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dong, W.; Shi-Sheng, Z. Color Image Recognition Method Based on the Prewitt Operator. In Proceedings of the International Conference on Computer Science & Software Engineering, Hubei, China, 12–14 December 2008; pp. 170–173. [Google Scholar]
  2. Kanopoulos, N.; Vasanthavada, N.; Baker, R.L. Design of an image edge detection filter using the Sobel operator. IEEE J. Solid State Circuits 2002, 23, 358–367. [Google Scholar] [CrossRef]
  3. Li, E.-S.; Zhu, S.-L.; Zhu, B.-S.; Zhao, Y.; Xia, C.-G.; Song, L.-H.; Soc, I.C. An Adaptive Edge-detection Method Based on the Canny Operator. In Proceedings of the International Conference on Environmental Science & Information Application Technology, Wuhan, China, 4–5 July 2009; Volume 1, pp. 465–469. [Google Scholar]
  4. Wang, X. Laplacian operator-based edge detectors. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 886–890. [Google Scholar] [CrossRef]
  5. Otsu, N. A Threshold Selection Method from Gray—Scale Histograms. IEEE Trans. SMC 1979, 9, 62–66. [Google Scholar]
  6. Tien Sy, N.; Begot, S.; Duculty, F.; Avila, M.; IEEE. Free-form anisotropy: A new method for crack detection on pavement surface images. In Proceedings of the IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 1069–1072. [Google Scholar]
  7. Bhowmick, S.; Nagarajaiah, S.; Veeraraghavan, A. Vision and Deep Learning-Based Algorithms to Detect and Quantify Cracks on Concrete Surfaces from UAV Videos. Sensors 2020, 20, 6229. [Google Scholar] [CrossRef]
  8. Lee, J.; Kim, H.-S.; Kim, N.; Ryu, E.-M.; Kang, J.-W. Learning to Detect Cracks on Damaged Concrete Surfaces Using Two-Branched Convolutional Neural Network. Sensors 2019, 19, 4796. [Google Scholar] [CrossRef] [Green Version]
  9. Islam, M.M.M.; Kim, J.-M. Vision-Based Autonomous Crack Detection of Concrete Structures Using a Fully Convolutional Encoder—Decoder Network. Sensors 2019, 19, 4251. [Google Scholar] [CrossRef] [Green Version]
  10. Yang, X.; Li, H.; Yu, Y.; Luo, X.; Huang, T.; Yang, X. Automatic Pixel-Level Crack Detection and Measurement Using Fully Convolutional Network. Comput. Aided Civ. Infrastruct. Eng. 2018, 33, 1090–1109. [Google Scholar] [CrossRef]
  11. Gopalakrishnan, K.; Khaitan, S.K.; Choudhary, A.; Agrawal, A. Deep Convolutional Neural Networks with transfer learning for computer vision-based data-driven pavement distress detection. Constr. Build. Mater. 2017, 157, 322–330. [Google Scholar] [CrossRef]
  12. Cha, Y.-J.; Choi, W.; Buyukozturk, O. Deep Learning-Based Crack Damage Detection Using Convolutional Neural Networks. Comput. Aided Civ. Infrastruct. Eng. 2017, 32, 361–378. [Google Scholar] [CrossRef]
  13. Zhang, L.; Yang, F.; Zhang, Y.D.; Zhu, Y.J.; IEEE. Road crack detection using deep convolutional neural network. In Proceedings of the IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016; pp. 3708–3712. [Google Scholar]
  14. Charron, N.; McLaughlin, E.; Phillips, S.; Goorts, K.; Narasimhan, S.; Waslander, S.L. Automated Bridge Inspection Using Mobile Ground Robotics. J. Struct. Eng. 2019, 145, 04019137. [Google Scholar] [CrossRef]
  15. Zhang, D.; Zou, Q.; Lin, H.; Xu, X.; He, L.; Gui, R.; Li, Q. Automatic pavement defect detection using 3D laser profiling technology. Autom. Constr. 2018, 96, 350–365. [Google Scholar] [CrossRef]
  16. Ouyang, W.; Xu, B. Pavement cracking measurements using 3D laser-scan images. Meas. Sci. Technol. 2013, 24, 24. [Google Scholar] [CrossRef]
  17. Tsai, Y.-C.J.; Li, F. Critical Assessment of Detecting Asphalt Pavement Cracks under Different Lighting and Low Intensity Contrast Conditions Using Emerging 3D Laser Technology. J. Transp. Eng. 2012, 138, 649–656. [Google Scholar] [CrossRef]
  18. Wang, K.C.P. Elements of automated survey of pavements and a 3D methodology. J. Mod. Transp. 2011, 19, 51–57. [Google Scholar]
  19. Jian, L.; Quan, S.; Chen, K.; Hao, C.; Kuan, H.; Chen, X. 3D Large-Scale Point Cloud Semantic Segmentation Using Optimal Feature Description Vector Network: OFDV-Net. IEEE Access 2020, 8, 226285–226296. [Google Scholar] [CrossRef]
  20. Kim, M.-K.; Sohn, H.; Chang, C.-C. Localization and Quantification of Concrete Spalling Defects Using Terrestrial Laser Scanning. J. Comput. Civil Eng. 2015, 29, 04014086. [Google Scholar] [CrossRef] [Green Version]
  21. Li, B.; Wang, K.C.P.; Zhang, A.; Fei, Y.; Sollazzo, G. Automatic Segmentation and Enhancement of Pavement Cracks Based on 3D Pavement Images. J. Adv. Transp. 2019, 2019, 9. [Google Scholar] [CrossRef]
  22. Cho, S.; Park, S.; Cha, G.; Oh, T. Development of Image Processing for Crack Detection on Concrete Structures through Terrestrial Laser Scanning Associated with the Octree Structure. Appl. Sci. 2018, 8, 2373. [Google Scholar] [CrossRef] [Green Version]
  23. Hadavandsiri, Z.; Lichti, D.D.; Jahraus, A.; Jarron, D. Concrete Preliminary Damage Inspection by Classification of Terrestrial Laser Scanner Point Clouds through Systematic Threshold Definition. Isprs Int. J. Geo Inf. 2019, 8, 585. [Google Scholar] [CrossRef] [Green Version]
  24. Guldur, B.; Yan, Y.; Hajjar, J.F. Condition Assessment of Bridges Using Terrestrial Laser Scanners. In Proceedings of the Structures Congress, Portland, OR, USA, 23–25 April 2015; pp. 355–366. [Google Scholar]
  25. Yang, H.; Xu, X. Intelligent crack extraction based on terrestrial laser scanning measurement. Meas. Control 2020, 53, 416–426. [Google Scholar] [CrossRef] [Green Version]
  26. Guan, H.; Li, J.; Yu, Y.; Chapman, M.; Wang, H.; Wang, C.; Zhai, R. Iterative Tensor Voting for Pavement Crack Extraction Using Mobile Laser Scanning Data. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1527–1537. [Google Scholar] [CrossRef]
  27. Jiang, H.; Li, Q.; Jiao, Q.; Wang, X.; Wu, L. Extraction of Wall Cracks on Earthquake-Damaged Buildings Based on TLS Point Clouds. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 3088–3096. [Google Scholar] [CrossRef]
  28. Yu, Y.; Li, J.; Guan, H.; Wang, C.; IEEE. 3D Crack Skeleton Extraction from Mobile LiDAR Point Clouds. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Quebec, QC, Canada, 13–18 July 2014; Volume 2014, pp. 914–917. [Google Scholar]
  29. Zhong, M.; Sui, L.; Wang, Z.; Hu, D. Pavement Crack Detection from Mobile Laser Scanning Point Clouds Using a Time Grid. Sensors 2020, 20, 4198. [Google Scholar] [CrossRef] [PubMed]
  30. Dai, K.; Li, A.; Zhang, H.; Chen, S.-E.; Pan, Y. Surface damage quantification of postearthquake building based on terrestrial laser scan data. Struct. Control Health Monit. 2018, 25, 18. [Google Scholar] [CrossRef]
  31. Valenca, J.; Puente, I.; Julio, E.; Gonzalez-Jorge, H.; Arias-Sanchez, P. Assessment of cracks on concrete bridges using image processing supported by laser scanning survey. Constr. Build. Mater. 2017, 146, 668–678. [Google Scholar] [CrossRef]
  32. Huang, J.; Liu, W.; Sun, X. A Pavement Crack Detection Method Combining 2D with 3D Information Based on Dempster-Shafer Theory. Comput. Aided Civ. Infrastruct. Eng. 2014, 29, 299–313. [Google Scholar] [CrossRef]
  33. Duda, R.O.; Hart, P.E. Use of Hough Transformation to Detect Lines and Curves in Pictures. Commun. ACM 1972, 15, 11–15. [Google Scholar] [CrossRef]
  34. Zou, Q.; Zhang, Z.; Li, Q.; Qi, X.; Wang, Q.; Wang, S. Deepcrack: Learning hierarchical convolutional features for crack detection. IEEE Trans. Image Process. 2018, 28, 1498–1512. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the proposed crack-detection method.
Figure 1. The flowchart of the proposed crack-detection method.
Sensors 21 01581 g001
Figure 2. Point clouds projection grid.
Figure 2. Point clouds projection grid.
Sensors 21 01581 g002
Figure 3. Bilinear interpolation.
Figure 3. Bilinear interpolation.
Sensors 21 01581 g003
Figure 4. Registration result of depth and gray image. (a) image; (b) depth image; (c) registration result.
Figure 4. Registration result of depth and gray image. (a) image; (b) depth image; (c) registration result.
Sensors 21 01581 g004
Figure 5. Comparison of different images. (a) Original 2D image; (b) depth image; (c) fusion image.
Figure 5. Comparison of different images. (a) Original 2D image; (b) depth image; (c) fusion image.
Sensors 21 01581 g005
Figure 6. Rough detection of cracks based on improved Otsu’s method. (a) Improved overall flow chart of Otsu’s crack detection method; (b) classification and segmentation of image sub-blocks.
Figure 6. Rough detection of cracks based on improved Otsu’s method. (a) Improved overall flow chart of Otsu’s crack detection method; (b) classification and segmentation of image sub-blocks.
Sensors 21 01581 g006
Figure 7. Comparison of crack detection effect of fusion image between traditional Otsu’s method and improved Otsu’s method. (a) Fusion image; (b) traditional Otsu’s method; (c) improved Otsu’s method.
Figure 7. Comparison of crack detection effect of fusion image between traditional Otsu’s method and improved Otsu’s method. (a) Fusion image; (b) traditional Otsu’s method; (c) improved Otsu’s method.
Sensors 21 01581 g007
Figure 8. Denoising and connection of cracks.
Figure 8. Denoising and connection of cracks.
Sensors 21 01581 g008
Figure 9. Schematic diagram of the effect before and after denoising and connection of crack. (a) Before denoising and connection of crack; (b) after denoising and connection of crack.
Figure 9. Schematic diagram of the effect before and after denoising and connection of crack. (a) Before denoising and connection of crack; (b) after denoising and connection of crack.
Sensors 21 01581 g009
Figure 10. Comparison before and after denoising and connection of cracks: (a) before crack denoising and connection of cracks; (b) after denoising and connection of cracks.
Figure 10. Comparison before and after denoising and connection of cracks: (a) before crack denoising and connection of cracks; (b) after denoising and connection of cracks.
Sensors 21 01581 g010
Figure 11. Comparison of crack detection results of different methods based on handheld point clouds: (a) image; (b) depth image; (c) manual annotation result; (d) 2D image detection result; (e) 3D point clouds detection result; (f) detection result of the proposed method.
Figure 11. Comparison of crack detection results of different methods based on handheld point clouds: (a) image; (b) depth image; (c) manual annotation result; (d) 2D image detection result; (e) 3D point clouds detection result; (f) detection result of the proposed method.
Sensors 21 01581 g011
Figure 12. Comparison of crack detection results of different methods based on terrestrial laser point clouds: (a) image; (b) depth image; (c) manual annotation result; (d) 2D image detection result; (e) 3D point clouds detection result; (f) detection result of proposed method.
Figure 12. Comparison of crack detection results of different methods based on terrestrial laser point clouds: (a) image; (b) depth image; (c) manual annotation result; (d) 2D image detection result; (e) 3D point clouds detection result; (f) detection result of proposed method.
Sensors 21 01581 g012aSensors 21 01581 g012b
Figure 13. Comparison of detection results between deep learning and the proposed method: (a) image; (b) DeepCrack detection result; (c) detection result of proposed method; (d) manual annotation result.
Figure 13. Comparison of detection results between deep learning and the proposed method: (a) image; (b) DeepCrack detection result; (c) detection result of proposed method; (d) manual annotation result.
Sensors 21 01581 g013
Table 1. The terrestrial laser scanner parameters.
Table 1. The terrestrial laser scanner parameters.
Sensor PictureTechnical SpecificationRIEGL VZ-2000
Sensors 21 01581 i001Scanning distance2000 m
Positioning Precision5 mm/100 m
Maximum angular resolution0.0015°
RangeHorizontal 360° (max)
Vertical 100° (max)
Table 2. Handheld laser scanner parameters.
Table 2. Handheld laser scanner parameters.
Sensor PictureTechnical SpecificationHandySCAN700
Sensors 21 01581 i002Positioning accuracy0.03 mm
Grid resolution0.2 mm
Scanning area275 × 250 mm
Table 3. Relevant parameters of the proposed method.
Table 3. Relevant parameters of the proposed method.
Scanner TypeCrackWeighting FactorDistanceCrack Basic Information
kikr
Handheld laser10.450.550.25 mWide crack, many stains
20.260.740.25 mNarrow crack, many stains
30.350.650.25 mNarrow crack, many stains
Terrestrial laser40.800.203.15 mNarrow crack, many stains
50.450.553.01 mWide crack, many stains
60.250.753.38 mNarrow crack, many stains
Table 4. Comparison of quantitative results of crack detection (handheld laser).
Table 4. Comparison of quantitative results of crack detection (handheld laser).
Evaluation IndexCrack2D Only3D Only2D and 3D
P (%)158.598.497.3
244.681.677.7
338.790.690.5
R (%)199.488.591.3
287.274.382.4
380.868.181.2
F1 (%)173.693.2 94.2
259.077.8 80.0
352.477.7 85.6
Table 5. Comparison of quantitative results of crack detection (terrestrial laser).
Table 5. Comparison of quantitative results of crack detection (terrestrial laser).
Evaluation IndexCrack2D Only3D Only2D and 3D
P (%)488.454.687.9
587.499.996.0
642.285.584.6
R (%)485.975.488.4
566.646.881.7
693.375.283.8
F1 (%)487.163.4 88.1
575.663.7 88.3
658.280.0 84.2
Table 6. Comparison of the average detection results of crack detection using different methods.
Table 6. Comparison of the average detection results of crack detection using different methods.
Method P ¯   ( % ) R ¯   ( % ) F 1 ¯   ( % )
2D only60.085.567.6
3D only85.171.476.0
2D and 3D89.084.886.7
Table 7. Comparison of crack detection results between DeepCrack and the proposed method.
Table 7. Comparison of crack detection results between DeepCrack and the proposed method.
MethodP (%)R (%)F1 (%)
Proposed method77.782.480.0
Deep Crack68.938.949.7
Table 8. Comparison of average results of crack detection by different scanners.
Table 8. Comparison of average results of crack detection by different scanners.
Scanner TypeMethod P ¯   ( % ) R ¯   ( % ) F 1 ¯   ( % )
NO2D only62.786.469.1
Handheld laser3D only90.276.982.9
2D and 3D88.585.086.6
NO2D only80.582.879.5
Terrestrial laser3D only80.065.869.0
2D and 3D89.584.686.9
Table 9. Image ground sampling distance (GSD) information.
Table 9. Image ground sampling distance (GSD) information.
Scanner TypeCrackTarget Diameter (cm)Target Radius (pixel)GSD (cm/pixel)
Handheld laser10.659.130.0051
20.648.270.0062
30.646.940.0064
Terrestrial laser4559.250.0422
5563.610.0393
6569.710.0359
Table 10. Comparison of the maximum width error between the proposed method and manual annotation.
Table 10. Comparison of the maximum width error between the proposed method and manual annotation.
Scanner TypeCrackDetection Width (cm)True Width (cm)Absolute (Errorcm)Relative Error (%)
Handheld laser12.06 2.23 0.16 7.29
21.02 0.920.12 10.56
31.40 1.58 0.18 11.34
Terrestrial laser42.66 2.70 0.04 1.56
52.00 2.12 0.12 5.56
65.52 5.63 0.11 1.91
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, X.; Li, J.; Huang, S.; Cui, H.; Liu, P.; Sun, Q. An Automatic Concrete Crack-Detection Method Fusing Point Clouds and Images Based on Improved Otsu’s Algorithm. Sensors 2021, 21, 1581. https://doi.org/10.3390/s21051581

AMA Style

Chen X, Li J, Huang S, Cui H, Liu P, Sun Q. An Automatic Concrete Crack-Detection Method Fusing Point Clouds and Images Based on Improved Otsu’s Algorithm. Sensors. 2021; 21(5):1581. https://doi.org/10.3390/s21051581

Chicago/Turabian Style

Chen, Xiaolong, Jian Li, Shuowen Huang, Hao Cui, Peirong Liu, and Quan Sun. 2021. "An Automatic Concrete Crack-Detection Method Fusing Point Clouds and Images Based on Improved Otsu’s Algorithm" Sensors 21, no. 5: 1581. https://doi.org/10.3390/s21051581

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop