Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Risk Assessment and Experimental Light-Balloon Deployment of a Stratospheric Vertical VLF Transmitter
Next Article in Special Issue
Assessment of Various Multimodal Fusion Approaches Using Synthetic Aperture Radar (SAR) and Electro-Optical (EO) Imagery for Vehicle Classification via Neural Networks
Previous Article in Journal
New Insights into Stroke from Continuous Passively Collected Temperature and Sleep Data Using Wrist-Worn Wearables
Previous Article in Special Issue
Super-Resolution Reconstruction Method of Pavement Crack Images Based on an Improved Generative Adversarial Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Irregular Scene Text Detection Based on a Graph Convolutional Network

1
College of Science, Jiangxi University of Science and Technology, Ganzhou 341000, China
2
College of Information Science and Engineering, Jiaxing University, Jiaxing 314001, China
3
Key Laboratory of Medical Electronics and Digital Health of Zhejiang Province, Jiaxing University, Jiaxing 314001, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(3), 1070; https://doi.org/10.3390/s23031070
Submission received: 25 December 2022 / Revised: 11 January 2023 / Accepted: 12 January 2023 / Published: 17 January 2023

Abstract

:
Detecting irregular or arbitrary shape text in natural scene images is a challenging task that has recently attracted considerable attention from research communities. However, limited by the CNN receptive field, these methods cannot directly capture relations between distant component regions by local convolutional operators. In this paper, we propose a novel method that can effectively and robustly detect irregular text in natural scene images. First, we employ a fully convolutional network architecture based on VGG16_BN to generate text components via the estimated character center points, which can ensure a high text component detection recall rate and fewer noncharacter text components. Second, text line grouping is treated as a problem of inferring the adjacency relations of text components with a graph convolution network (GCN). Finally, to evaluate our algorithm, we compare it with other existing algorithms by performing experiments on three public datasets: ICDAR2013, CTW-1500 and MSRA-TD500. The results show that the proposed method handles irregular scene text well and that it achieves promising results on these three public datasets.

1. Introduction

In recent years, with the popularity of smartphones and intelligent terminals, there has been a growing demand for the extraction of text information from scene images (such as those from intelligent transportation systems [1], geographic information systems [2], and automatic scene understanding [3]). In response, scene image text detection has attracted the attention of many researchers, and many text detection algorithms have been proposed, such as CTPN [4], FCEnet [5], and Quadbox [6]. However, the detection of irregular scene text is still an unsolved problem.
As methods based on connected regions are not affected by the lengths and shapes of text lines, they can achieve better performance in irregular scene text detection cases, as shown for CTPN [4] and TextSnake [7]. This kind of method usually employs convolutional neural network (CNN)-based approaches to predict the adjacency relations between text components. However, scene image texts are unfixed or irregular in terms of their positions, sizes, and orientations, and their underlying structures are non-Euclidean structures. When using a CNN to process such data, they need to be artificially converted to Euclidean-structured data, which can reduce the final performance, meaning these methods cannot directly capture relations between distant component regions due to the limitation of local convolutional operators [8]. On the other hand, graph convolutional networks have obvious advantages in processing non-Euclidean structured data. DRRG [8] detects irregular text via a graph convolutional network. This method generates components in each text region and then uses these text components to build a connected graph, where each node in the connected graph represents a text component. Then, a graph convolutional network is used to infer the adjacency relations between the nodes in the connected graph to generate text lines. However, this method produces a large number of noncharacter text components during the generation process, which affects the detection speed and performance of the approach. Inspired by DRRG [8], we propose an irregular scene text detection method based on a graph convolutional network. The main flowchart of our method is shown in Figure 1. During the text component extraction stage, to reduce the number of noncharacter text components, the estimated character center point is utilized after detecting the text regions with a fully convolutional network containing VGG16_BN as the backbone. In the text line grouping stage, we consider the process of grouping irregular text lines as a problem of inferring the adjacency relations of text components with a graph convolution network. Our paper makes two primary contributions.
(1) Different from the literature [8], we combine the text regions created by a fully convolutional network architecture based on VGG16_BN and the estimated character center points to generate text components, which can obviously reduce the number of noncharacter text components and help to improve the performance of text line grouping.
(2) We argue that the text line grouping problem can be posed as distinguishing edges that are adjacent and nonadjacent to an undirected graph. A relation inference network based on a graph convolutional network is designed to infer the adjacency relations of text components, in which the adjacency relationship between text components can be obtained independently of the spatial distance between text components.
The remainder of this paper is organized as follows. Section 2 presents an overview of previously published methods. Section 3 introduces the proposed method, and Section 4 reports the results of experiments conducted using the proposed system on the ICDAR2013, MSRA-TD500, and CTW-1500 datasets. Finally, conclusions and suggestions for future work are provided in Section 5.

2. Related Work

In this section, we summarize recent progress in arbitrary or irregular shape text detection in scene images. Further information about text detection in scenes is available in [1,9]. In recent years, detecting multioriented, arbitrary, or irregular text has become a popular topic in the domain of text detection in scenes, and numerous studies [5,6,8,10,11,12,13,14,15] concerning irregular text detection have been published. These methods can be roughly divided into the following three categories.
Regression-based methods. These methods treat the problem of scene text detection as a regression problem with text bounding boxes [14,16,17]. Textboxes [17] utilizes a sliding window and a convolution kernel for long rectangular candidate boxes to adapt to the shapes of text lines, but this method can only process horizontal texts and cannot work well when encountering arbitrarily shaped text. ABCnet [14] uses eight Bezier curve control points to describe arbitrarily shaped text, but it requires high-precision point controls, which seriously affects the final results. EAST [18] was designed to yield fast and accurate text detection results in natural scenes. Zhang et al. [16] proposed a novel adaptive boundary proposal network for arbitrary shape text detection, which can learn to directly produce accurate boundaries for arbitrary shape text without any postprocessing. Although these methods have achieved good performance in horizontal and multioriented text detection cases, they may fail to detect scene texts with large aspect ratios and directions.
Segmentation-based methods. These methods locate text instances based on pixel-level classification [11,12,19,20]. They usually employ deep convolutional neural networks to obtain the text segmentation region, and then several postprocessing steps are used to obtain the final text boxes. PilexLink [12] separates texts lying close to each other by predicting the pixel connections between different text instances. PSEnet [11] was proposed as a progressive scale expansion postprocessing algorithm, yielding greatly improved detection accuracy. Tian et al. [19] proposed a pixel embedding algorithm that calculates the feature distances between pixels and groups the pixels in the segmentation results according to these feature distances. These methods have obvious advantages when dealing with the segmentation of text and nontext data. However, during the training process with the text segmentation region, some irrelevant noncharacters may also be marked as characters. This may result in text line adhesion in the segmentation results of these methods, which can affect their segmentation performance.
Connected component-based methods. These methods usually detect individual text parts or characters first and then link or group the detected individual text parts or characters into final text instances by a postprocessing procedure. Because they have a more flexible representation and adapt well to irregularly shaped text, these methods are popular in arbitrarily shaped text detection [4,6,7,8,12,13]. PixelLink [12] separates texts lying close to each other by predicting pixel connections between different text instances. TextSnake [7] used ordered disks and text centerlines to model text instances, which made it able to represent text in arbitrary shapes. However, TextSnake still needs time-consuming and complicated postprocessing steps (centralizing, striding, and sliding) during inference. DRRG [8] also proposed a text detection method, in which each text instance is constructed by a series of ordered rectangular components(including text components and nontext components). These methods based on individual text parts generally divide the text regions into many parts (text parts and nontext parts), and many nontext parts are generated simultaneously and are fed to the following steps, which may increase the complexity and difficulty of calculation. Wei et al. [10] proposed a multioriented text detection method that generated character candidates by exhaustive segmentation. CRAFT [13] predicted character region maps and affinity maps by weakly supervised learning. As the text components contain only character regions, these methods can decrease the number of text components fed to subsequent steps and reduce the complexity and difficulty of calculations.
Relational inference. The connected component-based methods are usually robust for the lengths and shapes of text lines, but the performance of these methods mainly depend on the following text line grouping. PixelLink [12] used embedding features to provide instance information and generate the text area. In CRAFT [13], after predicting affinity maps by weakly supervised learning, the affinity map was used to group characters into a single instance. However, limited by the CNN receptive field, these methods cannot directly capture relations between distant component regions by local convolutional operators. To solve this problem, Zhang et al. [8] employed a graph convolutional network (GCN) to further reason and deduce the likelihood of linkages between the component and its neighbors based on local graphs. Their method achieved state-of-the-art performance on publicly available datasets. Zhang et al. [16] adopted an adaptive boundary deformation model combined with GCN and RNN to perform iterative boundary deformation to obtain a more accurate text instance shape. Their method achieved impressive results on challenging text-in-the-wild datasets such as TotalText [21].

3. Proposed Method

An overview of the overall architecture of our method is shown in Figure 1, and the whole framework is mainly divided into text component extraction, local inference graph establishment, deep adjacency relation inference, and text line generation. First, the text regions are extracted by the feature extraction network, and then the character text components are obtained according to the character center points in the text regions. Second, according to the basic properties of character text components, the local inference graph is established, which contains the basic connection relations between the character text components. Based on the local inference graph, the deep relational inference network further performs reasoning on the connection relationships between the character text components. Finally, the reasoning results are divided into different text instances according to the division of the connected regions.

3.1. Text Component Extraction

Text component extraction is an important preprocessing step whose main goal is to precisely generate text components, which are used for the subsequent text line grouping process. Different from other methods, we employ a two-stream convolutional neural network framework, which combines text region generation with character center point estimation to generate text components. This step can reduce the number of noncharacter text components without dropping the recall rate of the text component detection results.
Text region generation. To obtain the text region (TR), we use feature pyramid networks with VGG16 as the backbone to extract the feature maps of the given image. Finally, the feature maps are processed by classification and regression to obtain the text region. The architecture of the feature extraction network is shown in Figure 2.
Character center point estimation. To more accurately locate text components, we refer to the CRAFT method [13]. A fully convolutional network with VGG16_BN as the backbone is used to obtain a heatmap, which represents the probability that the pixel is the center of the character area. Then, the character center points are estimated based on the heatmaps. The workflow of character center point estimation is shown in Figure 3. First, the connected regions are obtained by threshold segmentation from the heatmap, and then the center points of the characters are obtained by the Otsu algorithm. The character center points will be directly used in the subsequent text component generation. The restriction of the character center point can ensure that the extracted text components are character text components and that the character text components are more representative in terms of feature attributes. On the other hand, the restriction of the character center point reduces the generation of nontext components, which can reduce the computational complexity.
Text component extraction. The subsequent task is text component generation. Different from DRRG, we combine the text region with character center point estimation to generate the text components, which can reduce the number of noncharacter text components. The processing flow of this step is shown in Figure 3. First, we divide the character center points into several groups according to the number of text regions, and the character center points in each group are sorted according to their X coordinate values. Second, we draw a straight line between two adjacent character center points in each group. As shown in Figure 4c, let L i represent the straight line that passes through p i and p i + 1 , and let K i be perpendicular to the line L i passing through the character point p i . Let t i and b i represent the intersection points between line K i and the boundaries of region R 1 . Then, the associated text component is represented by the rotated rectangle whose height (denoted by H i ) is the distance from point t i to point b i . The width (denoted by W i ) of the rotated rectangle is one-third of its height. According to this method, each character center point corresponds to a text component. More details are shown in Algorithm 1.
Algorithm 1: Text component extraction
Input: Character center point set P = { p 1 , p 2 p n }; Text regions R = { R 1 , R 2 R m }.
Output: Text components T C  = { T C 1 , T C 2 T C n }
1:
S←NULL
2:
for i = 1:m do
3:
       G i ←NULL
4:
      for each p in P but not in S do
5:
            if p is in R i  then
6:
                 add p to G i
7:
                 add p to S
8:
for i = 1:m do
9:
      The points in G i are sorted by their X coordinate values
10:
for i = 1:m do
11:
      for each p in G i  do
12:
            L p ←A line connecting node p to the next node
13:
            K p ←A line perpendicular to L p and passing through node p
14:
            t p , b p ←The boundary of R i intersects two points of K p
15:
            H p ← distance( t p , b p )
16:
            W p H p /3
17:
            θ p ← The angle between K p and the X-axis
18:
            T C p ← ( X p , Y p , H p , W p , θ p )
19:
           add T C p to T C
20:
return   T C
Detection loss. The text component extraction loss consists of two losses and is computed as
L = L c l s + L r e g
where L c l s is a smooth L1 [22] regression loss and L r e g is a cross-entropy classification loss. The classification loss L c l s is computed as
L c l s = L t r + λ 1 L t p + λ 2 L t n
where L t r represents the loss for the T R ; L t p only calculates pixels inside the T R , and L t n only calculates the pixels outside T R . In scene images, the text instances usually occupy a small area. Thus, the numbers of text pixels and nontext pixels are rather imbalanced. To make the network training process focus more on pixels that are hard to distinguish, we adopt the OHEM [23] strategy for the T R loss, in which the ratio between the negatives and positives is set to 3:1. In our experiments, the weights λ 1 and λ 2 are empirically set to 1.0 and 0.5, respectively. The regression loss L r is computed as
L r e g = L H + L θ
L H = s m o o t h L 1 1 n i = 1 n ( H i ˜ H i )
L θ = s m o o t h L 1 1 n i = 1 n ( θ i ˜ θ i )
where H i and θ i are ground-truth values; H i ˜ and θ i ˜ are the corresponding predicted values; H is the height of the text component in the ground truth; and n is the number of text components.

3.2. Local Inference Graph Establishment

After extracting the text components, the next task is to group the text components into text lines. We treat text line grouping as a problem of inferring the adjacency relations between text components with a graph convolution network. Each node in the graph corresponds to a text component. However, if all nodes on the fully connected graph are directly used for inference, the training time and difficulty will be increased. To solve this problem, DRRG [8] uses a method that generates a local inference graph, which consists of the pivot node and its first-order and second-order neighbor nodes. The first-order neighbor nodes consist of the 8 nearest neighbor nodes of the pivot, and the second-order neighbor nodes consist of the 4 nearest neighbors of the first-order neighbor nodes. Different from DRRG, our method only needs to consider four neighbor nodes, and we select the pivot and four first-order adjacent nodes and two second-order adjacent nodes to build the local inference graph, which can reduce the number of nodes participating in the reasoning calculation. More details about the local inference graph establishment process are shown in Figure 5. In our method, adjacency is judged by the similarity between nodes, and the similarity E s between the pivot node p and another node q is defined as:
E s = 1 E p q / max ( H , W )
E p q = ( X p X q ) 2 + ( Y p Y q ) 2
where H and W are the height and width of the corresponding images, respectively. E p q is the Euclidean distance between node p and node q.

3.3. Deep Adjacency Relation Inference

The local inference graph contains the basic adjacency relations between the text component nodes. However, simple link mapping or embedding mapping cannot fully reflect the adjacency relations between text component nodes, and we thus employ a deep relational inference network based on a GCN to further infer the adjacency relations between the text component nodes. In the deep adjacency relation inference process, we only need to infer the adjacency relations between the pivot and the first-order adjacent nodes, and we believe that the features of a node can also be affected by the features of its surrounding nodes, so the second-order adjacent nodes are used to provide fusion features for the first-order adjacent nodes. The inputs of the GCN usually include the feature matrix and adjacency matrix (denoted by X and A, respectively). The computational method of the two matrices is as follows:
Feature matrix X . The text components within the same text instance have similar geometric features, and each text component is made up of a rotated rectangle, so we combine the deep features with the geometric features as features for the text components. After extracting to a text component, we map its features to the RROI-Align layer and then obtain its deep features. Next, we obtain the geometric features of the text component through its five geometric attributes(X,Y,W,H, θ ). According to the literature [8,24,25], the geometric attributes of text components are embedded in high dimensional spaces to obtain geometric features. The formulas for the embedding calculation are shown in Equations (8) and (9). Finally, we concatenate the deep features and geometric features to obtain the feature matrix X of the text components.
ε 2 i ( z ) = c o s ( z 1000 2 i / C ε ) , i ( 0 , C ε / 2 i 1 )
ε 2 i + 1 ( z ) = s i n ( z 1000 2 i / C ε ) , i ( 0 , C ε / 2 i 1 )
Adjacency matrix A . We extract the adjacency matrix A based on the preliminary connection relations between the text component nodes in the local inference graph. For each local inference graph, if the text component node i is connected with text component node j, let A ( i , j ) = 1, otherwise, let A ( i , j ) = 0. We do not need to explore the adjacency relation between the node and itself, and define A ( i , i ) = 0.
Graph convolutional network. After the feature matrix X and the adjacency matrix A of the local inference graph are obtained, we employ the deep relation inference network based on a GCN to further infer the adjacency relations between the pivot and its first-order adjacent text component nodes. F k represents the feature matrix of the output at layer k, and the convolution layer is defined as follows:
F k = σ ( ( X k G X k ) W k )
G = D ˜ 1 / 2 A ˜ D ˜ 1 / 2
D ˜ i , i = j A ˜ i , j
where X k R N × d i n , F k R N × d o u t , and d i n , d o u t are the feature dimensions of the input nodes and output nodes, respectively; Λ is the diagonal matrix and N is the number of text components in the local inference graph; G represents the symmetric normalized Laplacian of size N × N; the ⊕ operator denotes concatenation along the feature matrix; W k is the trainable weight matrix of layer k; and σ represents a nonlinear activation function. We only back-propagate the gradients for the nodes on the 1-order neighbors in training, because we only care about the linkage between a pivot and its first-order neighbors. For testing, we also only consider the classification of 1-order nodes.

3.4. Text Line Generation

After completing the deep adjacency relation inference, the adjacency probability matrix (denoted by S) is obtained by summarizing the probability information of all local inference graphs. Let T H represent the threshold of the adjacency probability. If the adjacency probability between node i and node j is greater than the threshold T H , the edge between them is preserved (we also say that node i connects node j) and we let S ˜ ( i , j ) = 1 , otherwise S ˜ ( i , j ) = 0 . Then, we use breadth-first search (BFS) to find the connected subgraphs, and the set of subgraphs is represented by L = { L 1 , L 2 L k }. Finally, the nodes in each subgraph are sorted, and each element in the set L represents a final text line that we need to find.

4. Experiments

4.1. Datasets and Evaluation Methods

ICDAR2013 dataset. The ICDAR2013 dataset is inherited from the benchmark used in ICDAR 2011. Several images that are duplicated over training and testing sets of the ICDAR 2011 dataset are removed. In addition, a small portion of the ground-truth annotations is revised. A total of 229 images are used for training, and 233 images are used for testing. However, this dataset is a collection of natural images having horizontal and near-horizontal text appearances.
MSRA-TD500 dataset. The MSRA Text Detection 500 Database (MSRA-TD500) contains 500 natural images, which are taken from indoor (office and mall) and outdoor (street) scenes using a pocket camera. The indoor images are mainly signs, doorplates, and caution signs, while the outdoor images are mostly guide boards and billboards in complex backgrounds. The resolutions of the images vary from 1296 × 864 to 1920 × 1280. The text in this dataset may be in different languages (Chinese, English, or a mixture of both), fonts, sizes, colors, and orientations.
CTW-1500 dataset. The SCUT-CTW1500 dataset contains 1500 images: 1000 for training and 500 for testing. In particular, it provides 10,751 cropped text instance images, including 3530 with curved text. The images are manually harvested from the internet, image libraries such as Google Open-Image, or phone cameras. The dataset contains much horizontal and multioriented text.
To evaluate the detection performance of the proposed method on the above three datasets, the evaluation method proposed by Wolf and Jolion [26] is adopted to compare the detection performance of our method with that of other methods through precision (P), recall (R), and F-measure (F).

4.2. Implementation Details

The backbone of our network is the VGG16 model pretrained on ImageNet [27]. In the training phase, the input image size is set to 640 × 640, and data augmentations (resizing, flipping, rotation, cropping, padding, etc.) are also employed to produce better-performing models. During the training process, stochastic gradient descent (SGD) is selected as the optimizer of the model, and the initial learning rate is set to 0.001. Then, the learning rate decreases by 0.0001 every 100 epochs. All experiments are carried out on a computer equipped with an RTX3080 GPU graphics card and the Linux16.04 system, and the experimental environment includes Python3.6 and PyTorch1.7.

4.3. Ablation Study

4.3.1. Ablation Study on Text Component Extraction

To verify the effectiveness of our method in the text component generation stage, we conduct a text component ablation experiment on the MSRA-TD500 and CTW-1500 datasets. Our method is improved on the basis of DRRG, and we compare the experimental results with those of DRRG. We visualize the generated text components of the two methods, and the comparison results are shown in Figure 6. We also statistically compare the numbers of text components and detection results of the two methods. Through the data analysis comparison in Table 1, we find that the number of text components per image and the detection time decrease significantly, with the number of text components decreasing by 32% and 26%, and the detection time decreasing by 40% and 51%. Compared with those of DRRG, the F measures of our method on MSRA-TD500 and CTW-1500 are improved by 2.3% and 3.5%, respectively. The experimental results show that our text component generation method can both decrease the number of noncharacter text components and also improve the performance of text detection.

4.3.2. Ablation Experiment on Deep Relation Inference

Methods that directly detect text regions through feature extraction networks often encounter difficulties when dealing with segmentation between text lines. For example, two text regions are easy to segment into one region, such as a and b in Figure 7, or one region can be divided into two regions, such as c, d, e, and f in Figure 7. The relation inference network can achieve improved text region segmentation through the adjacency relationships of the text components and then produce better detection results. To verify the effectiveness of the adjacency inference network, we also conduct deep relation inference ablation experiments on the MSRA-TD500 and CTW-1500 datasets. It can be found from Table 2 that the precision, recall, and F-measure are increased by 6.5%, 6.6%, and 6.6%, respectively, on MSRA-TD500. On CTW-1500, the precision, recall, and F-measure are increased by 3.6%, 4.8%, and 4.3%, respectively. The improvement of these performance indices also proves the effectiveness of our proposed adjacency inference network.

4.4. Experimental Results and Discussion

Experiments on the ICDAR2013 dataset. We conducted experiments on the ICDAR2013 dataset, and some examples of the experimental results are shown in Figure 8. The results show that the method in this paper achieves a good detection effect on this dataset. The comparison between the detection results of our method and other text detection methods is shown in Table 3. The data in Table 3 show that our method achieves 92.8% precision, an 87.1% recall rate, and an 89.9% F-measure in the detection performance evaluation. The recall rate and F-measure of our method are better than those of the other methods.
Experiments on the MSRA-TD500 dataset. To verify the effectiveness of our method on a multilanguage scene text dataset, we also carry out experiments on the MSRA-TD500 dataset. Some examples of the experimental results are shown in Figure 9, and a comparison with the results of other methods is shown in Table 4. Table 4 shows that our method also achieves good results on the MSRA-TD500 dataset, with precision, recall rate, and F-measure values of 89.7%, 85.1%, and 87.4%, respectively. Compared with other methods, our method achieves the best recall rate and F-measure value.
Experiments on CTW-1500 dataset. In addition, to verify the robustness of our method in terms of detecting irregular scene text, we also select the CTW-1500 irregular and multidirectional text dataset for experiments. Some examples of the experimental results are shown in Figure 10, and the performance parameter comparison with other methods is shown in Table 5. Table 5 shows that the recall rate and F-measure of our method are better than those of other methods, reaching 85.4% and 86.1%.
Discussion. The experimental results show that after adding constraints, our method can reduce the generation rate of noncharacter text, which in turn reduces the detection rate of nontext and improves the recall rate. In addition, the deep relational inference network further modifies the text regions obtained by the FPN, and the text detection effect is further improved. In addition, the deep relational inference network further corrects the text regions obtained by the FPN, and the text detection effect is further improved. However, our method is not ideal when dealing with the detection of overlapping text (Figure 11a), low-resolution text (Figure 11b), and partially occluded text (Figure 11c). The main reason for this is that the character detection effect is impacted during the detection process.

4.5. Generalization Ability

To further verify the effectiveness of our method, we randomly collected some images of natural scenes containing irregular text from the internet and conducted experiments on the collected images. The experimental results are shown in Figure 12. As shown in Figure 12, the proposed method is robust to irregular scene texts, which shows that the proposed method has a good generalization ability. We believe that two main factors lead to the good results of our approach. On the one hand, we employ the connected component-based method, which is robust to the shapes and sizes of scene texts, to generate text commponents in text component stage. On the other hand, the GCN can obtain a better adjacency relation between two text component nodes during the text line grouping stage. The above shows that the generalization ability of our method is good.

5. Conclusions

This paper presents an irregular text detection method based on a graph convolution network. In the text component extraction stage, we combine character center point estimation with the text regions to extract more accurate text components. In the text line grouping stage, we treat text line grouping as a problem of inferring the adjacency relations of text components. From the experimental results, it is shown that our method is effective and robust in irregular scene text detection scenarios and achieves promising results on three public datasets. Future work will focus on two aspects.
(1) We will solve the limitations of the current methods, such as the detection of overlapping text, low-resolution text, and partially occluded text in scene text images.
(2) Combining our method with text recognition, we will design an end-to-end text detection and recognition approach.

Author Contributions

Conceptualization, S.Z., C.Z. and Y.W.; Writing–original draft, S.Z., C.Z. and Y.W.; Writing–review and editing, X.Z., Y.L. and L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported partially by grants from the Scientific Research Fund of the Zhejiang Provincial Education Department (No.Y202147449), the Zhejiang Public Welfare Technology Research Project Fund of China under Grant (No. LGG20F010010), and the Jiaxing Science and Technology Project (No. 2020AY10024 and No.2022AD10017).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Long, S.; He, X.; Yao, C. Scene text detection and recognition: The deep learning era. Int. J. Comput. Vis. 2021, 129, 161–184. [Google Scholar] [CrossRef]
  2. Yang, C.; Chen, M.; Xiong, Z.; Yuan, Y.; Wang, Q. Cm-net: Concentric mask based arbitrary-shaped text detection. IEEE Trans. Image Process. 2022, 31, 2864–2877. [Google Scholar] [CrossRef]
  3. Keserwani, P.; Saini, R.; Liwicki, M.; Roy, P.P. Robust Scene Text Detection for Partially Annotated Training Data. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 8635–8645. [Google Scholar] [CrossRef]
  4. Tian, Z.; Huang, W.; He, T.; He, P.; Qiao, Y. Detecting text in natural image with connectionist text proposal network. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 56–72. [Google Scholar]
  5. Zhu, Y.; Chen, J.; Liang, L.; Kuang, Z.; Zhang, W. Fourier Contour Embedding for Arbitrary-Shaped Text Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3123–3131. [Google Scholar]
  6. Keserwani, P.; Dhankhar, A.; Saini, R.; Roy, P.P. Quadbox: Quadrilateral Bounding Box Based Scene Text Detection Using Vector Regression. IEEE Access 2021, 9, 36802–36818. [Google Scholar] [CrossRef]
  7. Long, S.; Ruan, J.; Zhang, W.; He, X.; Wu, W.; Yao, C. TextSnake: A Flexible Representation for Detecting Text of Arbitrary Shapes. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  8. Zhang, S.X.; Zhu, X.; Hou, J.B.; Liu, C.; Yang, C.; Wang, H.; Yin, X.C. Deep relational reasoning graph network for arbitrary shape text detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9699–9708. [Google Scholar]
  9. Chen, X.; Jin, L.; Zhu, Y.; Luo, C.; Wang, T. Text recognition in the wild: A survey. ACM Comput. Surv. (CSUR) 2021, 54, 1–35. [Google Scholar] [CrossRef]
  10. Wei, Y.; Shen, W.; Zeng, D.; Ye, L.; Zhang, Z. Multi-oriented text detection from natural scene images based on a CNN and pruning non-adjacent graph edges. Signal Process. Image Commun. 2018, 64, 89–98. [Google Scholar] [CrossRef]
  11. Li, X.; Wang, W.; Hou, W.; Liu, R.Z.; Lu, T.; Yang, J. Shape Robust Text Detection with Progressive Scale Expansion Network. arXiv 2018, arXiv:1806.02559. [Google Scholar]
  12. Dan, D.; Liu, H.; Li, X.; Deng, C. PixelLink: Detecting Scene Text via Instance Segmentation. arXiv 2018, arXiv:1801.01315. [Google Scholar]
  13. Baek, Y.; Lee, B.; Han, D.; Yun, S.; Lee, H. Character Region Awareness for Text Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  14. Liu, Y.; Chen, H.; Shen, C.; He, T.; Jin, L.; Wang, L. Abcnet: Real-time scene text spotting with adaptive bezier-curve network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9809–9818. [Google Scholar]
  15. Wang, Y.; Mamat, H.; Xu, X.; Aysa, A.; Ubul, K. Scene Uyghur Text Detection Based on Fine-Grained Feature Representation. Sensors 2022, 22, 4372. [Google Scholar] [CrossRef]
  16. Zhang, S.X.; Zhu, X.; Yang, C.; Wang, H.; Yin, X.C. Adaptive boundary proposal network for arbitrary shape text detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual Conference, 19–25 June 2021; pp. 1305–1314. [Google Scholar]
  17. Liao, M.; Shi, B.; Bai, X.; Wang, X.; Liu, W. TextBoxes: A Fast Text Detector with a Single Deep Neural Network. arXiv 2016, arXiv:1611.06779. [Google Scholar] [CrossRef]
  18. Zhou, X.; Yao, C.; Wen, H.; Wang, Y.; Zhou, S.; He, W.; Liang, J. EAST: An Efficient and Accurate Scene Text Detector. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  19. Tian, Z.; Shu, M.; Lyu, P.; Li, R.; Zhou, C.; Shen, X.; Jia, J. Learning Shape-Aware Embedding for Scene Text Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  20. Zhang, S.X.; Zhu, X.; Hou, J.B.; Yang, C.; Yin, X.C. Kernel Proposal Network for Arbitrary Shape Text Detection. IEEE Trans. Neural Netw. Learn. Syst. 2022; Early Access. [Google Scholar] [CrossRef]
  21. Ch’ng, C.K.; Chan, C.S.; Liu, C.L. Total-text: Toward orientation robustness in scene text detection. Int. J. Doc. Anal. Recognit. (IJDAR) 2020, 23, 31–52. [Google Scholar] [CrossRef]
  22. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 1–9. [Google Scholar] [CrossRef] [Green Version]
  23. Shrivastava, A.; Gupta, A.; Girshick, R. Training region-based object detectors with online hard example mining. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 761–769. [Google Scholar]
  24. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  25. Gu, J.; Hu, H.; Wang, L.; Wei, Y.; Dai, J. Learning region features for object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 381–395. [Google Scholar]
  26. Wolf, C.; Jolion, J.M. Object count/area graphs for the evaluation of object detection and segmentation algorithms. Int. J. Doc. Anal. Recognit. 2006, 8, 280–296. [Google Scholar] [CrossRef]
  27. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  28. Wei, Y.; Zhang, Z.; Shen, W.; Zeng, D.; Fang, M.; Zhou, S. Text detection in scene images based on exhaustive segmentation. Signal Process. Image Commun. Publ. Eur. Assoc. Signal Process. 2017, 50, 1–8. [Google Scholar] [CrossRef]
  29. Gao, J.; Wang, Q.; Yuan, Y. Convolutional Regression Network for Multi-oriented Text Detection. IEEE Access 2019, 7, 96424–96433. [Google Scholar] [CrossRef]
  30. Jeon, M.; Jeong, Y.S. Compact and accurate scene text detector. Appl. Sci. 2020, 10, 2096. [Google Scholar] [CrossRef] [Green Version]
  31. Tang, J.; Yang, Z.; Wang, Y.; Zheng, Q.; Bai, X. Detecting Dense and Arbitrary-shaped Scene Text by Instance-aware Component Grouping. Pattern Recognit. 2019, 96, 106954. [Google Scholar] [CrossRef]
  32. Xu, Y.; Wang, Y.; Zhou, W.; Wang, Y.; Bai, X. TextField: Learning A Deep Direction Field for Irregular Scene Text Detection. IEEE Trans. Image Process. 2019, 28, 5566–5579. [Google Scholar] [CrossRef] [Green Version]
  33. Wan, Q.; Ji, H.; Shen, L. Self-attention based text knowledge mining for text detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5983–5992. [Google Scholar]
  34. Wang, X.; Jiang, Y.; Luo, Z.; Liu, C.L.; Choi, H.; Kim, S. Arbitrary Shape Scene Text Detection with Adaptive Text Region Representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  35. Wang, X.; Zheng, S.; Zhang, C.; Li, R.; Gui, L. R-YOLO: A real-time text detector for natural scenes with arbitrary rotation. Sensors 2021, 21, 888. [Google Scholar] [CrossRef] [PubMed]
  36. Raisi, Z.; Naiel, M.A.; Younes, G.; Wardell, S.; Zelek, J.S. Transformer-based text detection in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3162–3171. [Google Scholar]
  37. Liao, M.; Zou, Z.; Wan, Z.; Yao, C.; Bai, X. Real-time scene text detection with differentiable binarization and adaptive scale fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 919–931. [Google Scholar] [CrossRef] [PubMed]
  38. Feng, W.; He, W.; Yin, F.; Zhang, X.Y.; Liu, C.L. TextDragon: An End-to-End Framework for Arbitrary Shaped Text Spotting. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  39. Wang, Y.; Xie, H.; Zha, Z.; Xing, M.; Fu, Z.; Zhang, Y. ContourNet: Taking a Further Step toward Accurate Arbitrary-shaped Scene Text Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
Figure 1. Overview of our overall architecture. Our network mainly consists of four parts: Text component extraction, local inference graph establishment, deep adjacency relation inference, and text line generation.
Figure 1. Overview of our overall architecture. Our network mainly consists of four parts: Text component extraction, local inference graph establishment, deep adjacency relation inference, and text line generation.
Sensors 23 01070 g001
Figure 2. Architecture of the feature extraction network, where C R represents the classification and regression operation for the text components. C R R H × W × 5 has 2 channels for the logistic classification of text regions and 3 channels for the logistic regression of h 1 , h 2 , and θ .
Figure 2. Architecture of the feature extraction network, where C R represents the classification and regression operation for the text components. C R R H × W × 5 has 2 channels for the logistic classification of text regions and 3 channels for the logistic regression of h 1 , h 2 , and θ .
Sensors 23 01070 g002
Figure 3. The workflow of character center node estimation. (a) Source image. (b) Heatmap. (c) Segmentation result. (d) Character center points.
Figure 3. The workflow of character center node estimation. (a) Source image. (b) Heatmap. (c) Segmentation result. (d) Character center points.
Sensors 23 01070 g003
Figure 4. The processing flow of text component generation. (a) The result of character center point grouping. (b) Text region R 1 and its character center points. (c) One of the text components.
Figure 4. The processing flow of text component generation. (a) The result of character center point grouping. (b) Text region R 1 and its character center points. (c) One of the text components.
Sensors 23 01070 g004
Figure 5. Local inference graph generation. (a) Source image. (b) Graph of text component nodes. (c) Local inference graph. “1st” indicates the first-order nodes and “2st” denotes the second-order nodes.
Figure 5. Local inference graph generation. (a) Source image. (b) Graph of text component nodes. (c) Local inference graph. “1st” indicates the first-order nodes and “2st” denotes the second-order nodes.
Sensors 23 01070 g005
Figure 6. Comparative examples of the text component extraction results of our method and DRRG. The first row contains the text component instances extracted by our method, and the second row contains text component instances extracted by DRRG.
Figure 6. Comparative examples of the text component extraction results of our method and DRRG. The first row contains the text component instances extracted by our method, and the second row contains text component instances extracted by DRRG.
Sensors 23 01070 g006
Figure 7. Experimental results of two detection methods. The images in the top row are the detection results obtained directly through the feature extraction network, and the images in the next row are the detection results obtained with the addition of the GCN-based relation inference network.
Figure 7. Experimental results of two detection methods. The images in the top row are the detection results obtained directly through the feature extraction network, and the images in the next row are the detection results obtained with the addition of the GCN-based relation inference network.
Sensors 23 01070 g007
Figure 8. Detection examples of the proposed method in the ICDAR2013 dataset.
Figure 8. Detection examples of the proposed method in the ICDAR2013 dataset.
Sensors 23 01070 g008
Figure 9. Detection examples obtained by the proposed method on the MSRA-TD500 dataset.
Figure 9. Detection examples obtained by the proposed method on the MSRA-TD500 dataset.
Sensors 23 01070 g009
Figure 10. Detection examples produced by the proposed method on the MSRA-TD500 dataset.
Figure 10. Detection examples produced by the proposed method on the MSRA-TD500 dataset.
Sensors 23 01070 g010
Figure 11. Some failed detection examples of the proposed method: (a) overlapping text; (b) low-resolution text; (c) partially occluded text. The red lines mark text boxes, and the text markers ①–⑨ represent the text that is un-detected.
Figure 11. Some failed detection examples of the proposed method: (a) overlapping text; (b) low-resolution text; (c) partially occluded text. The red lines mark text boxes, and the text markers ①–⑨ represent the text that is un-detected.
Sensors 23 01070 g011
Figure 12. Examples of internet and life scene text images.
Figure 12. Examples of internet and life scene text images.
Sensors 23 01070 g012
Table 1. Results of ablation experiments involving text component extraction.
Table 1. Results of ablation experiments involving text component extraction.
DatasetsMethodsAve_numAve_time(s)P(%)R(%)F(%)
MSRA-TD500DRRG410.25288.182.385.1
Ours280.15189.785.187.4
CTW-1500DRRG840.37483.881.582.6
Ours620.18386.785.486.1
Note: Ave_num is the average number of text components per image; Ave_time is the average detection time per image.
Table 2. Results of ablation experiments of deep relation inference.
Table 2. Results of ablation experiments of deep relation inference.
DatasetsMethodsP (%)R (%)F (%)
MSRA-TD500Baseline83.278.580.8
Baseline+GCN89.785.187.4
CTW-1500Baseline83.180.681.8
Baseline+GCN86.785.486.1
Note: Baseline is directly obtained by text region; Baseline+GCN is a relation inference network based on GCN added to Baseline.
Table 3. Experimental results of various methods on the ICDAR2013 dataset.
Table 3. Experimental results of various methods on the ICDAR2013 dataset.
MethodsP (%)R (%)F (%)
Wei Y et al. [28]83.577.280.2
Wei Y et al. [10]87.381.184.3
Gao et al. [29]90.080.085.0
CAST [30]94.069.080.0
Ours92.887.189.9
Table 4. Experimental results of various methods on the MSRA-TD500 dataset.
Table 4. Experimental results of various methods on the MSRA-TD500 dataset.
MethodsP (%)R (%)F (%)
SegLink [31]86.070.077.0
TextField [32]87.475.981.3
CRAFT [13]88.278.282.9
Wan et al. [33]81.677.279.3
Wang et al. [34]85.082.083.0
DRRG [8]88.182.385.1
R-YOLO [35]90.281.985.8
Zobeir Raisi et al. [36]90.983.887.2
Liao et al. [37]91.583.387.2
Ours89.785.187.4
Table 5. Experimental results of various methods on the CTW-1500 dataset.
Table 5. Experimental results of various methods on the CTW-1500 dataset.
MethodsP (%)R (%)F (%)
TextDragon [38]84.582.883.6
TextField [32]83.079.881.4
Zhang et al [16]87.780.684.0
CRAFT [13]86.081.183.5
Wan et al. [33]85.178.281.5
DRRG [8]85.983.084.4
CountourNet [39]83.784.183.9
PCENet [5]87.683.485.5
Ours86.785.486.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, S.; Zhou, C.; Li, Y.; Zhang, X.; Ye, L.; Wei, Y. Irregular Scene Text Detection Based on a Graph Convolutional Network. Sensors 2023, 23, 1070. https://doi.org/10.3390/s23031070

AMA Style

Zhang S, Zhou C, Li Y, Zhang X, Ye L, Wei Y. Irregular Scene Text Detection Based on a Graph Convolutional Network. Sensors. 2023; 23(3):1070. https://doi.org/10.3390/s23031070

Chicago/Turabian Style

Zhang, Shiyu, Caiying Zhou, Yonggang Li, Xianchao Zhang, Lihua Ye, and Yuanwang Wei. 2023. "Irregular Scene Text Detection Based on a Graph Convolutional Network" Sensors 23, no. 3: 1070. https://doi.org/10.3390/s23031070

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop